var/home/core/zuul-output/0000755000175000017500000000000015114126515014526 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114142704015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004761514215114142545017711 0ustar rootrootDec 03 21:41:42 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 21:41:42 crc restorecon[4560]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:42 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 21:41:43 crc restorecon[4560]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 21:41:43 crc kubenswrapper[4715]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 21:41:43 crc kubenswrapper[4715]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 21:41:43 crc kubenswrapper[4715]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 21:41:43 crc kubenswrapper[4715]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 21:41:43 crc kubenswrapper[4715]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 21:41:43 crc kubenswrapper[4715]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.435539 4715 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439161 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439184 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439191 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439198 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439203 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439210 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439216 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439223 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439231 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439239 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439245 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439251 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439256 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439276 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439282 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439287 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439293 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439298 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439306 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439313 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439318 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439324 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439330 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439335 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439341 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439347 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439352 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439358 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439364 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439369 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439375 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439382 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439388 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439394 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439401 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439407 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439412 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439417 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439423 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439429 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439434 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439441 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439447 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439453 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439459 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439465 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439471 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439476 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439483 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439490 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439495 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439526 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439532 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439538 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439544 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439549 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439554 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439559 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439565 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439571 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439576 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439581 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439586 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439592 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439597 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439603 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439608 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439613 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439618 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439625 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.439632 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.439965 4715 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.439984 4715 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.439995 4715 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440004 4715 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440012 4715 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440019 4715 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440027 4715 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440047 4715 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440054 4715 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440060 4715 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440067 4715 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440073 4715 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440079 4715 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440085 4715 flags.go:64] FLAG: --cgroup-root="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440092 4715 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440098 4715 flags.go:64] FLAG: --client-ca-file="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440104 4715 flags.go:64] FLAG: --cloud-config="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440109 4715 flags.go:64] FLAG: --cloud-provider="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440115 4715 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440127 4715 flags.go:64] FLAG: --cluster-domain="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440133 4715 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440140 4715 flags.go:64] FLAG: --config-dir="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440145 4715 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440152 4715 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440160 4715 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440166 4715 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440172 4715 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440178 4715 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440184 4715 flags.go:64] FLAG: --contention-profiling="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440190 4715 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440196 4715 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440203 4715 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440209 4715 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440219 4715 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440228 4715 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440235 4715 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440241 4715 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440248 4715 flags.go:64] FLAG: --enable-server="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440254 4715 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440266 4715 flags.go:64] FLAG: --event-burst="100" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440272 4715 flags.go:64] FLAG: --event-qps="50" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440287 4715 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440293 4715 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440312 4715 flags.go:64] FLAG: --eviction-hard="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440319 4715 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440325 4715 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440332 4715 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440338 4715 flags.go:64] FLAG: --eviction-soft="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440344 4715 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440350 4715 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440356 4715 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440362 4715 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440368 4715 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440374 4715 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440380 4715 flags.go:64] FLAG: --feature-gates="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440387 4715 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440393 4715 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440399 4715 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440405 4715 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440412 4715 flags.go:64] FLAG: --healthz-port="10248" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440418 4715 flags.go:64] FLAG: --help="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440423 4715 flags.go:64] FLAG: --hostname-override="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440429 4715 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440435 4715 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440441 4715 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440448 4715 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440454 4715 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440461 4715 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440466 4715 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440473 4715 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440479 4715 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440485 4715 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440491 4715 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440497 4715 flags.go:64] FLAG: --kube-reserved="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440526 4715 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440532 4715 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440538 4715 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440545 4715 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440551 4715 flags.go:64] FLAG: --lock-file="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440565 4715 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440573 4715 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440579 4715 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440588 4715 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440594 4715 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440600 4715 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440606 4715 flags.go:64] FLAG: --logging-format="text" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440612 4715 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440618 4715 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440624 4715 flags.go:64] FLAG: --manifest-url="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440630 4715 flags.go:64] FLAG: --manifest-url-header="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440638 4715 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440644 4715 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440651 4715 flags.go:64] FLAG: --max-pods="110" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440658 4715 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440664 4715 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440669 4715 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440675 4715 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440682 4715 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440689 4715 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440695 4715 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440709 4715 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440716 4715 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440722 4715 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440729 4715 flags.go:64] FLAG: --pod-cidr="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440735 4715 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440749 4715 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440754 4715 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440761 4715 flags.go:64] FLAG: --pods-per-core="0" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440767 4715 flags.go:64] FLAG: --port="10250" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440773 4715 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440779 4715 flags.go:64] FLAG: --provider-id="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440785 4715 flags.go:64] FLAG: --qos-reserved="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440791 4715 flags.go:64] FLAG: --read-only-port="10255" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440797 4715 flags.go:64] FLAG: --register-node="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440803 4715 flags.go:64] FLAG: --register-schedulable="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440817 4715 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440828 4715 flags.go:64] FLAG: --registry-burst="10" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440834 4715 flags.go:64] FLAG: --registry-qps="5" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440841 4715 flags.go:64] FLAG: --reserved-cpus="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440846 4715 flags.go:64] FLAG: --reserved-memory="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440855 4715 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440861 4715 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440868 4715 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440874 4715 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440881 4715 flags.go:64] FLAG: --runonce="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440887 4715 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440894 4715 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440901 4715 flags.go:64] FLAG: --seccomp-default="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440907 4715 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440913 4715 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440920 4715 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440927 4715 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440933 4715 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440941 4715 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440947 4715 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440953 4715 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440959 4715 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440966 4715 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440972 4715 flags.go:64] FLAG: --system-cgroups="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440978 4715 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440988 4715 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440994 4715 flags.go:64] FLAG: --tls-cert-file="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.440999 4715 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441012 4715 flags.go:64] FLAG: --tls-min-version="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441018 4715 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441024 4715 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441030 4715 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441038 4715 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441044 4715 flags.go:64] FLAG: --v="2" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441053 4715 flags.go:64] FLAG: --version="false" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441062 4715 flags.go:64] FLAG: --vmodule="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441077 4715 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441084 4715 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441286 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441294 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441301 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441306 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441313 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441319 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441324 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441330 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441336 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441341 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441346 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441351 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441358 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441363 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441368 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441374 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441379 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441384 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441390 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441395 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441401 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441406 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441411 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441417 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441423 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441428 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441434 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441439 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441445 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441450 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441455 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441460 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441466 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441471 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441486 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441491 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441497 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441527 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441533 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441538 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441544 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441549 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441554 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441559 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441564 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441571 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441578 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441584 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441590 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441596 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441601 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441608 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441614 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441621 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441628 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441634 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441639 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441645 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441651 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441657 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441662 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441668 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441673 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441678 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441684 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441689 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441694 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441699 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441705 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441712 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.441729 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.441748 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.460284 4715 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.460358 4715 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460557 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460574 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460586 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460597 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460610 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460621 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460633 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460645 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460655 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460666 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460675 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460685 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460693 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460746 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460758 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460768 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460778 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460788 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460799 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460809 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460819 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460828 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460837 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460846 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460855 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460864 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460873 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460883 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460892 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460901 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460910 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460918 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460928 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460937 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460945 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460954 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460962 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460970 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460980 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460988 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.460997 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461005 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461014 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461023 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461031 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461039 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461048 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461056 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461065 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461074 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461086 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461100 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461110 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461121 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461131 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461140 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461150 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461159 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461167 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461180 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461189 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461198 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461208 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461218 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461227 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461236 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461245 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461254 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461263 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461275 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461286 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.461301 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461619 4715 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461637 4715 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461647 4715 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461659 4715 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461671 4715 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461680 4715 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461688 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461697 4715 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461706 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461716 4715 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461724 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461732 4715 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461744 4715 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461758 4715 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461771 4715 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461780 4715 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461791 4715 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461800 4715 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461810 4715 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461818 4715 feature_gate.go:330] unrecognized feature gate: Example Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461828 4715 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461840 4715 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461850 4715 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461860 4715 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461868 4715 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461877 4715 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461885 4715 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461894 4715 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461903 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461911 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461919 4715 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461928 4715 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461936 4715 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461945 4715 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461953 4715 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461961 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461970 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461978 4715 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461986 4715 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.461996 4715 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462004 4715 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462013 4715 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462021 4715 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462030 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462039 4715 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462048 4715 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462057 4715 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462067 4715 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462075 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462084 4715 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462092 4715 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462100 4715 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462109 4715 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462118 4715 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462126 4715 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462135 4715 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462146 4715 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462157 4715 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462166 4715 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462176 4715 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462186 4715 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462195 4715 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462204 4715 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462212 4715 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462220 4715 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462230 4715 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462238 4715 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462247 4715 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462255 4715 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462263 4715 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.462272 4715 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.462287 4715 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.463177 4715 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.468322 4715 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.468469 4715 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.469349 4715 server.go:997] "Starting client certificate rotation" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.469392 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.469685 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-24 18:24:54.308873821 +0000 UTC Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.469814 4715 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 500h43m10.839062927s for next certificate rotation Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.478266 4715 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.481604 4715 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.492617 4715 log.go:25] "Validated CRI v1 runtime API" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.509647 4715 log.go:25] "Validated CRI v1 image API" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.511739 4715 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.515762 4715 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-21-35-05-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.515809 4715 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:40 fsType:tmpfs blockSize:0}] Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.545004 4715 manager.go:217] Machine: {Timestamp:2025-12-03 21:41:43.542483336 +0000 UTC m=+0.285194021 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:14ababa8-2b39-4191-882f-3b54d73698cb BootID:cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:40 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:2e:e9:93 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:2e:e9:93 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:a7:d1:5b Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:c7:ef:cc Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ce:dd:56 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:13:e6:f5 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:96:d3:3d:ef:08:4e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:aa:05:a8:35:2d:4f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.547988 4715 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.549256 4715 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.550531 4715 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.550943 4715 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.551019 4715 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.551443 4715 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.551466 4715 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.551786 4715 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.551849 4715 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.552166 4715 state_mem.go:36] "Initialized new in-memory state store" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.552935 4715 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.554089 4715 kubelet.go:418] "Attempting to sync node with API server" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.554132 4715 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.554182 4715 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.554208 4715 kubelet.go:324] "Adding apiserver pod source" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.554227 4715 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.557188 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.557189 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.557310 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.557336 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.557951 4715 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.558377 4715 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559298 4715 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559881 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559907 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559916 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559925 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559938 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559957 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559969 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559983 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.559993 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.560001 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.560014 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.560022 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.560405 4715 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.560837 4715 server.go:1280] "Started kubelet" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.561084 4715 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.561369 4715 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.562048 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.562463 4715 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 21:41:43 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.563348 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.563371 4715 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.563803 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 10:52:22.713581626 +0000 UTC Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.564116 4715 server.go:460] "Adding debug handlers to kubelet server" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.564317 4715 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.564339 4715 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.564450 4715 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.564183 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187dd2865f067be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 21:41:43.560813541 +0000 UTC m=+0.303524136,LastTimestamp:2025-12-03 21:41:43.560813541 +0000 UTC m=+0.303524136,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.564523 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.564991 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.569805 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.569879 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.575977 4715 factory.go:55] Registering systemd factory Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.576011 4715 factory.go:221] Registration of the systemd container factory successfully Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.576415 4715 factory.go:153] Registering CRI-O factory Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.576457 4715 factory.go:221] Registration of the crio container factory successfully Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.576602 4715 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.576645 4715 factory.go:103] Registering Raw factory Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.576677 4715 manager.go:1196] Started watching for new ooms in manager Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.579102 4715 manager.go:319] Starting recovery of all containers Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590272 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590369 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590386 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590403 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590422 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590440 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590457 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590473 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590492 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590585 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.590601 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591587 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591671 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591722 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591747 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591772 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591827 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591847 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591888 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591911 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591930 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591950 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591971 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.591990 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592009 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592027 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592052 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592075 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592093 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592112 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592131 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592152 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592170 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592186 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592204 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592223 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592240 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592261 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592280 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592298 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592318 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592338 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592362 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592411 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592429 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592447 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592464 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592486 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592527 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592548 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592567 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592583 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592607 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592626 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592646 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592664 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592682 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592703 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592724 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592741 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592760 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592782 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592801 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592869 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592889 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592905 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592922 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592938 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592957 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592975 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.592992 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593010 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593027 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593044 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593062 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593079 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593125 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593195 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593219 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593272 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593295 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593316 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593335 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593353 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593373 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593390 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593407 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593424 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593441 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593459 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593477 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593494 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593538 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593555 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593571 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593589 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593607 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593623 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593640 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593656 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593671 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593690 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593707 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593723 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593749 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593769 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593787 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.593807 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594135 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594155 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594172 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594191 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594212 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594227 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594244 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594261 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594278 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594294 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594310 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594326 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594342 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594357 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594376 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594392 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594411 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594427 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594444 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594464 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594479 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594494 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594532 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594547 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594565 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594582 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594598 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.594615 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.595796 4715 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.595889 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.595919 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.595947 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.595972 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.595994 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596018 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596047 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596075 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596101 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596123 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596146 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596169 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596193 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596217 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596239 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596261 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596282 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596303 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596325 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596346 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596369 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596391 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596414 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596436 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596464 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596486 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596539 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596564 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596592 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596614 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596638 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596665 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596686 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596710 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596732 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596757 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596779 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596804 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596828 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596851 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596872 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596892 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596912 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596932 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596952 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596973 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.596995 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597016 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597037 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597057 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597078 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597188 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597210 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597230 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597254 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597276 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597298 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597318 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597340 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597360 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597381 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597401 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597425 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597446 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597466 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597487 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597532 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597552 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597573 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597594 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597617 4715 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597640 4715 reconstruct.go:97] "Volume reconstruction finished" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.597654 4715 reconciler.go:26] "Reconciler: start to sync state" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.607250 4715 manager.go:324] Recovery completed Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.621847 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.623913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.623972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.623989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.625152 4715 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.625184 4715 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.625223 4715 state_mem.go:36] "Initialized new in-memory state store" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.629400 4715 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.632887 4715 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.632954 4715 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.633007 4715 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.633089 4715 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 21:41:43 crc kubenswrapper[4715]: W1203 21:41:43.633845 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.633961 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.649158 4715 policy_none.go:49] "None policy: Start" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.650536 4715 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.650573 4715 state_mem.go:35] "Initializing new in-memory state store" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.664711 4715 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.730885 4715 manager.go:334] "Starting Device Plugin manager" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.731200 4715 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.731222 4715 server.go:79] "Starting device plugin registration server" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.731738 4715 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.731769 4715 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.732129 4715 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.732223 4715 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.732232 4715 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.734677 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.734865 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.736126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.736174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.736186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.736351 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.736923 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.736963 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737483 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737622 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737659 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.737972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.739366 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.739373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.739428 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.739444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.739525 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.739578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.739598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.739856 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.740000 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.740045 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741629 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741822 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.741943 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.742711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.742834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.742894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.743001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.743104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.743191 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.743583 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.743830 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.746123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.746472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.746558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.766281 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.809655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.809736 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.809783 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.809824 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.809858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.809925 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.809969 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.809998 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.810027 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.810071 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.810106 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.810150 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.810190 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.810220 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.810277 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.831989 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.833635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.833672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.833684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.833715 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 21:41:43 crc kubenswrapper[4715]: E1203 21:41:43.834248 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911646 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911714 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911747 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911777 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911856 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911949 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912156 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.911955 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912018 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912004 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912041 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912089 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912559 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912222 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912674 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912746 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912662 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912730 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912914 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.912994 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.913072 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.913100 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.913147 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:43 crc kubenswrapper[4715]: I1203 21:41:43.913177 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.035218 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.037363 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.037401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.037415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.037447 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 21:41:44 crc kubenswrapper[4715]: E1203 21:41:44.037994 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.080102 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.089729 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 21:41:44 crc kubenswrapper[4715]: W1203 21:41:44.115585 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-6a3efbb4eca02fbba134679fc8a44b35cee35772adeae1f7b0ddb83b663ae8bc WatchSource:0}: Error finding container 6a3efbb4eca02fbba134679fc8a44b35cee35772adeae1f7b0ddb83b663ae8bc: Status 404 returned error can't find the container with id 6a3efbb4eca02fbba134679fc8a44b35cee35772adeae1f7b0ddb83b663ae8bc Dec 03 21:41:44 crc kubenswrapper[4715]: W1203 21:41:44.118749 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-fe2320ea7eeae3da2dc53ef5270af0bc8d19e69fde4354b4bf3c374a0bdb193c WatchSource:0}: Error finding container fe2320ea7eeae3da2dc53ef5270af0bc8d19e69fde4354b4bf3c374a0bdb193c: Status 404 returned error can't find the container with id fe2320ea7eeae3da2dc53ef5270af0bc8d19e69fde4354b4bf3c374a0bdb193c Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.122788 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.132678 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.138904 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:44 crc kubenswrapper[4715]: W1203 21:41:44.154489 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-eacdc99f37cdb3c4ef59e5850f528770ebe3ae71931c5ec784aeb1b819cf24ec WatchSource:0}: Error finding container eacdc99f37cdb3c4ef59e5850f528770ebe3ae71931c5ec784aeb1b819cf24ec: Status 404 returned error can't find the container with id eacdc99f37cdb3c4ef59e5850f528770ebe3ae71931c5ec784aeb1b819cf24ec Dec 03 21:41:44 crc kubenswrapper[4715]: W1203 21:41:44.167316 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-3c37b8cce3dc300b61a79480b8509839a31b1524cfcdbc4add451fe409d31d66 WatchSource:0}: Error finding container 3c37b8cce3dc300b61a79480b8509839a31b1524cfcdbc4add451fe409d31d66: Status 404 returned error can't find the container with id 3c37b8cce3dc300b61a79480b8509839a31b1524cfcdbc4add451fe409d31d66 Dec 03 21:41:44 crc kubenswrapper[4715]: E1203 21:41:44.167413 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.438414 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.440266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.440293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.440301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:44 crc kubenswrapper[4715]: W1203 21:41:44.440257 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.440340 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 21:41:44 crc kubenswrapper[4715]: E1203 21:41:44.440386 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 03 21:41:44 crc kubenswrapper[4715]: E1203 21:41:44.441038 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.563189 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.564229 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 21:05:21.086879553 +0000 UTC Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.564302 4715 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 887h23m36.522580728s for next certificate rotation Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.639631 4715 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073" exitCode=0 Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.639713 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.639807 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3c37b8cce3dc300b61a79480b8509839a31b1524cfcdbc4add451fe409d31d66"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.639903 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.641761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.641819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.641840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.643567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.643590 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"eacdc99f37cdb3c4ef59e5850f528770ebe3ae71931c5ec784aeb1b819cf24ec"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.645820 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b" exitCode=0 Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.645893 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.645910 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8d837cea121e70a407d62eef9bd27c7084c780e4ba0bfe46d5c11a1dcb8acc81"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.646170 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.647736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.647792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.647814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.648682 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf" exitCode=0 Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.648736 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.648753 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6a3efbb4eca02fbba134679fc8a44b35cee35772adeae1f7b0ddb83b663ae8bc"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.648865 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.649912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.649934 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.649943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.651822 4715 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae" exitCode=0 Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.651844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.651861 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fe2320ea7eeae3da2dc53ef5270af0bc8d19e69fde4354b4bf3c374a0bdb193c"} Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.651910 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.652586 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.652635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.652645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.653063 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.653752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.653789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:44 crc kubenswrapper[4715]: I1203 21:41:44.653807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:44 crc kubenswrapper[4715]: W1203 21:41:44.675752 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:44 crc kubenswrapper[4715]: E1203 21:41:44.675892 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 03 21:41:44 crc kubenswrapper[4715]: W1203 21:41:44.860381 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:44 crc kubenswrapper[4715]: E1203 21:41:44.860650 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 03 21:41:44 crc kubenswrapper[4715]: E1203 21:41:44.968700 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Dec 03 21:41:45 crc kubenswrapper[4715]: W1203 21:41:45.140745 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Dec 03 21:41:45 crc kubenswrapper[4715]: E1203 21:41:45.140845 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.243117 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.245637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.245679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.245690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.245722 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.658424 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b212fc5927f3c43b22e22d9f3e58bc44022ee097b55548ce38d99f12341f1e74"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.658644 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.660411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.660537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.660584 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.664104 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.664146 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.664158 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.664257 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.665202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.665253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.665271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.668792 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.668848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.668820 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.668873 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.670092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.670118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.670128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.672685 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.672733 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.672755 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.672774 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.675725 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a" exitCode=0 Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.675783 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a"} Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.675932 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.677224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.677278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:45 crc kubenswrapper[4715]: I1203 21:41:45.677290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.683464 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5"} Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.683616 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.688929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.689108 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.689141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.690875 4715 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63" exitCode=0 Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.690979 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63"} Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.691054 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.691139 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.692178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.692225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.692244 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.692326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.692355 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:46 crc kubenswrapper[4715]: I1203 21:41:46.692370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:47 crc kubenswrapper[4715]: I1203 21:41:47.697700 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8"} Dec 03 21:41:47 crc kubenswrapper[4715]: I1203 21:41:47.697789 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef"} Dec 03 21:41:47 crc kubenswrapper[4715]: I1203 21:41:47.697821 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70"} Dec 03 21:41:47 crc kubenswrapper[4715]: I1203 21:41:47.697736 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 21:41:47 crc kubenswrapper[4715]: I1203 21:41:47.697901 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:47 crc kubenswrapper[4715]: I1203 21:41:47.699475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:47 crc kubenswrapper[4715]: I1203 21:41:47.699581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:47 crc kubenswrapper[4715]: I1203 21:41:47.699607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.584916 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.707977 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d"} Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.708090 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba"} Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.708137 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.709231 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.709329 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.709890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.709958 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.709986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.710628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.710681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:48 crc kubenswrapper[4715]: I1203 21:41:48.710699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.667759 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.667963 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.675673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.675739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.675758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.675877 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.710196 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.710357 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.711346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.711420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.711440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.711570 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.711596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.711608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:49 crc kubenswrapper[4715]: I1203 21:41:49.743039 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.285013 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.285273 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.287446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.287535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.287560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.331947 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.484233 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.484582 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.486074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.486132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.486152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.712978 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.713645 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.712996 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.715059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.715140 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.715161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.715700 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.715767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:50 crc kubenswrapper[4715]: I1203 21:41:50.715786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:51 crc kubenswrapper[4715]: I1203 21:41:51.786063 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 21:41:51 crc kubenswrapper[4715]: I1203 21:41:51.786296 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:51 crc kubenswrapper[4715]: I1203 21:41:51.787841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:51 crc kubenswrapper[4715]: I1203 21:41:51.787880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:51 crc kubenswrapper[4715]: I1203 21:41:51.787891 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:52 crc kubenswrapper[4715]: I1203 21:41:52.418564 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:52 crc kubenswrapper[4715]: I1203 21:41:52.418737 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 21:41:52 crc kubenswrapper[4715]: I1203 21:41:52.418792 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:52 crc kubenswrapper[4715]: I1203 21:41:52.420298 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:52 crc kubenswrapper[4715]: I1203 21:41:52.420341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:52 crc kubenswrapper[4715]: I1203 21:41:52.420352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:52 crc kubenswrapper[4715]: I1203 21:41:52.743719 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 21:41:52 crc kubenswrapper[4715]: I1203 21:41:52.743862 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.009419 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.009667 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.010781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.010833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.010851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.355364 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.355535 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.356771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.356862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:53 crc kubenswrapper[4715]: I1203 21:41:53.356889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:53 crc kubenswrapper[4715]: E1203 21:41:53.739743 4715 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 21:41:55 crc kubenswrapper[4715]: E1203 21:41:55.246694 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 03 21:41:55 crc kubenswrapper[4715]: E1203 21:41:55.477914 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": net/http: TLS handshake timeout" event="&Event{ObjectMeta:{crc.187dd2865f067be5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 21:41:43.560813541 +0000 UTC m=+0.303524136,LastTimestamp:2025-12-03 21:41:43.560813541 +0000 UTC m=+0.303524136,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 21:41:55 crc kubenswrapper[4715]: I1203 21:41:55.563727 4715 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 21:41:56 crc kubenswrapper[4715]: E1203 21:41:56.570257 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 03 21:41:56 crc kubenswrapper[4715]: W1203 21:41:56.753337 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.753441 4715 trace.go:236] Trace[265058851]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 21:41:46.751) (total time: 10001ms): Dec 03 21:41:56 crc kubenswrapper[4715]: Trace[265058851]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (21:41:56.753) Dec 03 21:41:56 crc kubenswrapper[4715]: Trace[265058851]: [10.001831363s] [10.001831363s] END Dec 03 21:41:56 crc kubenswrapper[4715]: E1203 21:41:56.753466 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 21:41:56 crc kubenswrapper[4715]: W1203 21:41:56.796880 4715 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.796990 4715 trace.go:236] Trace[685160911]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 21:41:46.794) (total time: 10002ms): Dec 03 21:41:56 crc kubenswrapper[4715]: Trace[685160911]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (21:41:56.796) Dec 03 21:41:56 crc kubenswrapper[4715]: Trace[685160911]: [10.002261127s] [10.002261127s] END Dec 03 21:41:56 crc kubenswrapper[4715]: E1203 21:41:56.797014 4715 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.847877 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.849394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.849436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.849449 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.849476 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.967518 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.967597 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.974345 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Dec 03 21:41:56 crc kubenswrapper[4715]: I1203 21:41:56.974421 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.594621 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.594894 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.596896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.596969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.596993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.601473 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.734785 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.736126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.736174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:41:58 crc kubenswrapper[4715]: I1203 21:41:58.736190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.821756 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.822027 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.823562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.823633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.823669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.840276 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.967601 4715 trace.go:236] Trace[1715266678]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 21:41:48.134) (total time: 13833ms): Dec 03 21:42:01 crc kubenswrapper[4715]: Trace[1715266678]: ---"Objects listed" error: 13833ms (21:42:01.967) Dec 03 21:42:01 crc kubenswrapper[4715]: Trace[1715266678]: [13.83335994s] [13.83335994s] END Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.967647 4715 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.968191 4715 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 21:42:01 crc kubenswrapper[4715]: E1203 21:42:01.969573 4715 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.969908 4715 trace.go:236] Trace[1150839517]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 21:41:47.581) (total time: 14388ms): Dec 03 21:42:01 crc kubenswrapper[4715]: Trace[1150839517]: ---"Objects listed" error: 14388ms (21:42:01.969) Dec 03 21:42:01 crc kubenswrapper[4715]: Trace[1150839517]: [14.388398203s] [14.388398203s] END Dec 03 21:42:01 crc kubenswrapper[4715]: I1203 21:42:01.969951 4715 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.023689 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:54220->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.023758 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:54220->192.168.126.11:17697: read: connection reset by peer" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.023776 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:54208->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.023871 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:54208->192.168.126.11:17697: read: connection reset by peer" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.023958 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.024390 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.024457 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.028695 4715 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.028757 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.031388 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.031710 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.227907 4715 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.565707 4715 apiserver.go:52] "Watching apiserver" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.568553 4715 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.568858 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c"] Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.569164 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.569684 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.569745 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.569784 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.569803 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.569838 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.569838 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.569907 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.570130 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.572328 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.572343 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.573968 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.574341 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.574569 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.574633 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.574747 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.574764 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.576463 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.605904 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.615987 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.630383 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.641810 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.654488 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.665550 4715 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.668477 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671227 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671287 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671324 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671358 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671389 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671419 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671465 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671495 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671553 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671584 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671616 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671646 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671678 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671749 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671781 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671810 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671867 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671912 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671948 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671980 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672012 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672041 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672137 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672167 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672199 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672228 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672258 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672286 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671580 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672320 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672351 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672380 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672409 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672439 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672467 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672495 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672587 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672620 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672681 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672732 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672769 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672798 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672830 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672871 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672912 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672993 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673026 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673058 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673121 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673152 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673182 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673235 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673266 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673300 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673335 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673418 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673453 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673484 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673541 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673575 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673607 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673639 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673669 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673699 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673730 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673761 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673793 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673824 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673857 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673890 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673926 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673966 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673997 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674029 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674062 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674093 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674124 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674157 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674188 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674218 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674249 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674279 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674309 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674343 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674377 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674410 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674442 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674474 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674532 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674564 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674596 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674626 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674657 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674691 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674722 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674755 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674786 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674817 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674848 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674881 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674913 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674945 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674978 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675012 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675044 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675077 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675143 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675176 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675211 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675245 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675278 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675309 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675431 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675468 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675538 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675573 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675607 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675642 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675675 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675709 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675744 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675776 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675808 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675843 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675875 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675907 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675938 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675971 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676029 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676095 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676150 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676266 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676315 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676362 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676411 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676459 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676714 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676769 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676805 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676847 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676880 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676916 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676949 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676980 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677016 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677049 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677082 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677116 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677149 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677184 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677219 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677252 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677317 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677351 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677410 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677448 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677493 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677575 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677615 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677648 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677682 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677718 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677751 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677787 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677820 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677853 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677887 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677921 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677962 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677999 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678036 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678069 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678115 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678291 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678345 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678392 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678433 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678485 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.671800 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678575 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672090 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672095 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678622 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678675 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678736 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678792 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678846 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678898 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678953 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679006 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679060 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679111 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679162 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679215 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679271 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679323 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679373 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679490 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679643 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679701 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679756 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679814 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679940 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680051 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680110 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680226 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680279 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680344 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680452 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680487 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680555 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680587 4715 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672260 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672285 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672555 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672646 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672666 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672680 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672708 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672891 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672919 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.672964 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673054 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673095 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673185 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673289 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673582 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673604 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673790 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.673951 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674024 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674112 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674180 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674220 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674310 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674414 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674430 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674444 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674492 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674701 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674855 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674941 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.674958 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675202 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675213 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675256 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675349 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675425 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675438 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675450 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675490 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675701 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.687574 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675932 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.675956 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676195 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.676320 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677066 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677282 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677397 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677589 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.677726 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678063 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.678101 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679867 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.679904 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680040 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.687662 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680332 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680259 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680668 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680967 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681056 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681087 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681205 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681318 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.688027 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681587 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681605 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681641 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681716 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.681824 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.682019 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.682801 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.682849 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.683306 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.683460 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.683438 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.683839 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.683980 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.684024 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.684066 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.684120 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.684306 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.684359 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.684785 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.685156 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.685157 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.685423 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.685539 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.685989 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.686023 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.686131 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.686731 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.687111 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.687156 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.680519 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.687994 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.688196 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.688394 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.688566 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.688642 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.688851 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.688861 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.689015 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.689200 4715 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.689332 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.689311 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.689780 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.690094 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.690216 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.690277 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.690381 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.690391 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.690788 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.691634 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.691146 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.691185 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.691659 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.691338 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.691576 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.691717 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.691759 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.692069 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.692283 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.692379 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.692682 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.693073 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.693388 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.693787 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.693948 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.694093 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.694381 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.694472 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.694610 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:03.194588058 +0000 UTC m=+19.937298733 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.694837 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.694914 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.694917 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.694960 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.695338 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.695641 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.695979 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.695999 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.696091 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.696520 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.696455 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.696821 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.696841 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.696954 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.697162 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.697191 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.697277 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.697661 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.697847 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.697914 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.697991 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:03.197973818 +0000 UTC m=+19.940684423 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.698828 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:42:03.198275605 +0000 UTC m=+19.940986230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.699534 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.704843 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.705030 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.705136 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.706304 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.706670 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.706734 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.706784 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.706997 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.707220 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.709582 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.711823 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.712063 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.712496 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.712553 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.712569 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.713982 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.714143 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.714179 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.714193 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:03.212612349 +0000 UTC m=+19.955323094 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.714431 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.714600 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.714678 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.714685 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.714812 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.714810 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.714866 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.714905 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.714933 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.715128 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.715540 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.715855 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:03.214998752 +0000 UTC m=+19.957709387 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.715860 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.715972 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.718376 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.718404 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.718433 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.718867 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.718905 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.719243 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.719395 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.719319 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.719607 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.719639 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.719712 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.720471 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.721146 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.721180 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.721318 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.721936 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.721989 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.722027 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.722211 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.722367 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.728723 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.731021 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.732009 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.737226 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.747271 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.749341 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5" exitCode=255 Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.749396 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5"} Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.753578 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:42:02 crc kubenswrapper[4715]: E1203 21:42:02.757767 4715 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.758797 4715 scope.go:117] "RemoveContainer" containerID="2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.759448 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.762100 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.762925 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.774996 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782048 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782144 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782373 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782410 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782427 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782444 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782460 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782474 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782493 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782574 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782593 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782684 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782705 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782720 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782738 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782754 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782770 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782786 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782803 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782819 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782835 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782854 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782870 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782887 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782904 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782920 4715 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782935 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782952 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782968 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782984 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.782999 4715 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783016 4715 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783033 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783049 4715 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783065 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783082 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783100 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783116 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783168 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783189 4715 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783208 4715 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783224 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783240 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783291 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783309 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783328 4715 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783346 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783364 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783383 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783401 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783417 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783434 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783453 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783494 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783548 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783565 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783582 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783598 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783614 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783631 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783648 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783665 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783709 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783725 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783741 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783758 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783776 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783794 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783811 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783828 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783834 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783847 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783845 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783909 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783925 4715 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783939 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783952 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783965 4715 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783977 4715 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.783990 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784003 4715 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784014 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784028 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784040 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784054 4715 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784068 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784079 4715 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784090 4715 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784103 4715 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784115 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784128 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784139 4715 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784152 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784164 4715 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784176 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784188 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784200 4715 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784212 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784223 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784259 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784308 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784320 4715 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784331 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784398 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784413 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784427 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784439 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784475 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784486 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784634 4715 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784650 4715 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784662 4715 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784675 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784714 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784725 4715 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784736 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784747 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784759 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784799 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784814 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784874 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.784910 4715 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785005 4715 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785042 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785054 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785068 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785081 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785092 4715 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785103 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785114 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785126 4715 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785138 4715 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785152 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785164 4715 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785177 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785190 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785204 4715 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785217 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785229 4715 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785242 4715 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785253 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785266 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785279 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785291 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785273 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785302 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785423 4715 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785436 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785450 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785461 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785494 4715 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785574 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785588 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785629 4715 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785644 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785655 4715 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785667 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785703 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785718 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785729 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785828 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785844 4715 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785890 4715 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785905 4715 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785921 4715 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785967 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785981 4715 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.785993 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786004 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786016 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786053 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786066 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786078 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786089 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786101 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786139 4715 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786151 4715 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786162 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786174 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786191 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786233 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786245 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786257 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786269 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786280 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786318 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786330 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786341 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786352 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786364 4715 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786404 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786416 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786428 4715 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786438 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.786472 4715 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.797484 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.809811 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.821973 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.833238 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.888137 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.894005 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 21:42:02 crc kubenswrapper[4715]: I1203 21:42:02.901166 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 21:42:02 crc kubenswrapper[4715]: W1203 21:42:02.904909 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-7881e068e2205b9b6c55028fcfbcc75c13b68a7fd01902fa22890a2ec34870de WatchSource:0}: Error finding container 7881e068e2205b9b6c55028fcfbcc75c13b68a7fd01902fa22890a2ec34870de: Status 404 returned error can't find the container with id 7881e068e2205b9b6c55028fcfbcc75c13b68a7fd01902fa22890a2ec34870de Dec 03 21:42:02 crc kubenswrapper[4715]: W1203 21:42:02.908159 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-7ce943ec69ab6faf1eed9b246d32176f31fc96c14e2a5d9af4f9f0f85758c7c8 WatchSource:0}: Error finding container 7ce943ec69ab6faf1eed9b246d32176f31fc96c14e2a5d9af4f9f0f85758c7c8: Status 404 returned error can't find the container with id 7ce943ec69ab6faf1eed9b246d32176f31fc96c14e2a5d9af4f9f0f85758c7c8 Dec 03 21:42:02 crc kubenswrapper[4715]: W1203 21:42:02.911462 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-72525be792a78d87772807ea95a1e80945b61998f70f177633e2a18c853e6b4e WatchSource:0}: Error finding container 72525be792a78d87772807ea95a1e80945b61998f70f177633e2a18c853e6b4e: Status 404 returned error can't find the container with id 72525be792a78d87772807ea95a1e80945b61998f70f177633e2a18c853e6b4e Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.096401 4715 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.291030 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291274 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:42:04.291222202 +0000 UTC m=+21.033932807 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.291138 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.291542 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.291687 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291484 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291740 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291754 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291819 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:04.291783907 +0000 UTC m=+21.034494512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291637 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291860 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:04.291852229 +0000 UTC m=+21.034562834 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.291899 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291970 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.291998 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:04.291989362 +0000 UTC m=+21.034699967 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.292100 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.292117 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.292128 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:03 crc kubenswrapper[4715]: E1203 21:42:03.292160 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:04.292150847 +0000 UTC m=+21.034861462 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.638234 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.638730 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.640020 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.640743 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.642426 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.642975 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.643604 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.644821 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.645408 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.646301 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.646846 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.647960 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.648494 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.649250 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.650371 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.651009 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.652631 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.653201 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.653945 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.655842 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.657929 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.664741 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.665481 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.665925 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.666562 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.666947 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.667547 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.668158 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.669992 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.670555 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.671357 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.671825 4715 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.671925 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.673701 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.674723 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.675204 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.677232 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.678413 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.678909 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.679524 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.680929 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.681921 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.683014 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.684020 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.685244 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.686094 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.687345 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.688236 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.689453 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.690548 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.691695 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.692193 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.693305 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.693854 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.694821 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.695655 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.697132 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.719235 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.731438 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.743878 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.756696 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472"} Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.756858 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add"} Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.756934 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7ce943ec69ab6faf1eed9b246d32176f31fc96c14e2a5d9af4f9f0f85758c7c8"} Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.757842 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed"} Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.758046 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7881e068e2205b9b6c55028fcfbcc75c13b68a7fd01902fa22890a2ec34870de"} Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.758721 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.759934 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.761181 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab"} Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.761805 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.763633 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"72525be792a78d87772807ea95a1e80945b61998f70f177633e2a18c853e6b4e"} Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.771924 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.789371 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.800640 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.810375 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.822612 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.835168 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.849418 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.865733 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.922214 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.944350 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:03 crc kubenswrapper[4715]: I1203 21:42:03.962650 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:04 crc kubenswrapper[4715]: I1203 21:42:04.300543 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:04 crc kubenswrapper[4715]: I1203 21:42:04.300608 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:04 crc kubenswrapper[4715]: I1203 21:42:04.300633 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:04 crc kubenswrapper[4715]: I1203 21:42:04.300651 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:04 crc kubenswrapper[4715]: I1203 21:42:04.300669 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300739 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:42:06.300697889 +0000 UTC m=+23.043408484 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300763 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300810 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300827 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300830 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300872 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:06.300863243 +0000 UTC m=+23.043573938 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300761 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300924 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:06.300902904 +0000 UTC m=+23.043613509 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300947 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:06.300937995 +0000 UTC m=+23.043648610 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300952 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.300996 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.301013 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.301075 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:06.301055258 +0000 UTC m=+23.043765863 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:04 crc kubenswrapper[4715]: I1203 21:42:04.634029 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:04 crc kubenswrapper[4715]: I1203 21:42:04.634056 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.634297 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:04 crc kubenswrapper[4715]: I1203 21:42:04.634367 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.634450 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:04 crc kubenswrapper[4715]: E1203 21:42:04.634607 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.169729 4715 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.171909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.171967 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.171985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.172068 4715 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.184149 4715 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.184259 4715 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.185684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.185738 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.185757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.185786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.185804 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: E1203 21:42:05.218853 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.223744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.223776 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.223787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.223804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.223816 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: E1203 21:42:05.242264 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.248618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.248827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.248862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.248893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.248914 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: E1203 21:42:05.268368 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.273538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.273587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.273603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.273626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.273642 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: E1203 21:42:05.299337 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.304836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.304885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.304901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.304924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.304962 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: E1203 21:42:05.326186 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: E1203 21:42:05.326455 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.328882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.328960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.328978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.329034 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.329053 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.432645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.432712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.432731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.432757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.432776 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.535752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.535828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.535851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.535911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.535936 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.638385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.638448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.638472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.638625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.638660 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.740606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.740651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.740664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.740681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.740693 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.769422 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5"} Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.796279 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.812713 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.826804 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.840876 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.842847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.842907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.842964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.842982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.842992 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.855025 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.873521 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.885487 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.897941 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.908317 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:05Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.945299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.945330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.945337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.945350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:05 crc kubenswrapper[4715]: I1203 21:42:05.945359 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:05Z","lastTransitionTime":"2025-12-03T21:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.047583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.047622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.047634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.047651 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.047660 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.150824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.150896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.150909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.150932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.150948 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.254982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.255680 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.255719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.255743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.255757 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.334985 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.335066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.335090 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.335111 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.335128 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335200 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335291 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:42:10.335231888 +0000 UTC m=+27.077942523 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335355 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:10.335339401 +0000 UTC m=+27.078050036 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335587 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335640 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335672 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335696 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335678 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:10.335662819 +0000 UTC m=+27.078373424 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335781 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:10.335759272 +0000 UTC m=+27.078470087 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335804 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335821 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.335833 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.336064 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:10.33605384 +0000 UTC m=+27.078764445 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.358314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.358354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.358368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.358388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.358401 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.461385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.461451 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.461468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.461493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.461529 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.565017 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.565079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.565091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.565128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.565139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.634234 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.634265 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.634336 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.634481 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.634361 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:06 crc kubenswrapper[4715]: E1203 21:42:06.634721 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.667693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.667760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.667779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.667804 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.667824 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.774997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.775111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.775137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.775171 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.775202 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.878427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.878476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.878494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.878547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.878566 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.981303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.981354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.981372 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.981397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:06 crc kubenswrapper[4715]: I1203 21:42:06.981415 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:06Z","lastTransitionTime":"2025-12-03T21:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.084128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.084199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.084225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.084254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.084276 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.187854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.188038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.188063 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.188092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.188112 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.291781 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.291863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.291886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.291921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.291941 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.395033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.395081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.395093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.395109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.395121 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.498389 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.498457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.498474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.498533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.498550 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.601642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.601701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.601716 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.601744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.601760 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.703867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.703926 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.703943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.703970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.703989 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.807380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.807453 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.807477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.807538 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.807566 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.911491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.912228 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.912248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.912272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:07 crc kubenswrapper[4715]: I1203 21:42:07.912290 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:07Z","lastTransitionTime":"2025-12-03T21:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.017401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.017434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.017442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.017457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.017466 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.119382 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.119450 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.119467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.119493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.119543 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.221598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.221663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.221677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.221720 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.221734 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.325120 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.325178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.325205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.325225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.325239 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.428715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.428743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.428752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.428765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.428774 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.530709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.530744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.530753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.530769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.530778 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.632938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.632987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.633003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.633021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.633035 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.633206 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.633235 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.633217 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:08 crc kubenswrapper[4715]: E1203 21:42:08.633403 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:08 crc kubenswrapper[4715]: E1203 21:42:08.633475 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:08 crc kubenswrapper[4715]: E1203 21:42:08.633580 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.735872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.735921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.735932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.735947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.735956 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.837672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.837714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.837731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.837754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.837771 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.939867 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.939914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.939923 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.939939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:08 crc kubenswrapper[4715]: I1203 21:42:08.939949 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:08Z","lastTransitionTime":"2025-12-03T21:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.041825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.041865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.041877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.041913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.041922 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.145291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.145336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.145345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.145360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.145370 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.248210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.248253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.248265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.248283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.248297 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.350707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.350753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.350767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.350787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.350801 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.353058 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-9wz2q"] Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.353348 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-wmhrj"] Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.353817 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9wz2q" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.355048 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2wcn9"] Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.355542 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.356150 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.357276 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.357302 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.358253 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.358425 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.358675 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.358832 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.360127 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.360195 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.360453 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.360973 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.367463 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.391930 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.403565 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.420072 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.439105 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.453322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.453392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.453406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.453422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.453433 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.465254 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468545 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-cni-bin\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468603 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-conf-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468685 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-os-release\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468719 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-os-release\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468776 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-socket-dir-parent\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468801 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/af42f462-12a2-4488-b408-8753ac62bcf4-hosts-file\") pod \"node-resolver-9wz2q\" (UID: \"af42f462-12a2-4488-b408-8753ac62bcf4\") " pod="openshift-dns/node-resolver-9wz2q" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468826 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-system-cni-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468847 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-netns\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468877 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-kubelet\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468906 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtwvp\" (UniqueName: \"kubernetes.io/projected/af42f462-12a2-4488-b408-8753ac62bcf4-kube-api-access-gtwvp\") pod \"node-resolver-9wz2q\" (UID: \"af42f462-12a2-4488-b408-8753ac62bcf4\") " pod="openshift-dns/node-resolver-9wz2q" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.468932 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-etc-kubernetes\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469097 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cnibin\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469125 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-daemon-config\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469235 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-k8s-cni-cncf-io\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-hostroot\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469326 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-multus-certs\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469395 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9174d88f-f7f0-4b0c-942f-39d027b97e34-cni-binary-copy\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469475 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-system-cni-dir\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469544 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469620 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-cni-multus\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469646 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b24hf\" (UniqueName: \"kubernetes.io/projected/9174d88f-f7f0-4b0c-942f-39d027b97e34-kube-api-access-b24hf\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469765 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzbc2\" (UniqueName: \"kubernetes.io/projected/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-kube-api-access-wzbc2\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.469830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cni-binary-copy\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.470121 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-cni-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.470154 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-cnibin\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.482642 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.512582 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.544396 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.555928 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.555989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.556000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.556022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.556033 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.563207 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571386 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-system-cni-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571439 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-netns\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571464 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-kubelet\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571490 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtwvp\" (UniqueName: \"kubernetes.io/projected/af42f462-12a2-4488-b408-8753ac62bcf4-kube-api-access-gtwvp\") pod \"node-resolver-9wz2q\" (UID: \"af42f462-12a2-4488-b408-8753ac62bcf4\") " pod="openshift-dns/node-resolver-9wz2q" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571535 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-etc-kubernetes\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571556 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cnibin\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571577 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-daemon-config\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571584 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-kubelet\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571596 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-netns\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571641 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-k8s-cni-cncf-io\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571646 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-etc-kubernetes\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571607 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-k8s-cni-cncf-io\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-system-cni-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571674 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cnibin\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-hostroot\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571884 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-hostroot\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571886 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-multus-certs\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571919 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-run-multus-certs\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.571989 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9174d88f-f7f0-4b0c-942f-39d027b97e34-cni-binary-copy\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572017 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-system-cni-dir\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572039 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572083 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-cni-multus\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572106 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b24hf\" (UniqueName: \"kubernetes.io/projected/9174d88f-f7f0-4b0c-942f-39d027b97e34-kube-api-access-b24hf\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572127 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzbc2\" (UniqueName: \"kubernetes.io/projected/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-kube-api-access-wzbc2\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572133 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-cni-multus\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572145 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cni-binary-copy\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-system-cni-dir\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-cni-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572188 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-cnibin\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572217 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-cni-bin\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572236 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-conf-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572252 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572267 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-os-release\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572285 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-os-release\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-socket-dir-parent\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572315 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/af42f462-12a2-4488-b408-8753ac62bcf4-hosts-file\") pod \"node-resolver-9wz2q\" (UID: \"af42f462-12a2-4488-b408-8753ac62bcf4\") " pod="openshift-dns/node-resolver-9wz2q" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572385 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/af42f462-12a2-4488-b408-8753ac62bcf4-hosts-file\") pod \"node-resolver-9wz2q\" (UID: \"af42f462-12a2-4488-b408-8753ac62bcf4\") " pod="openshift-dns/node-resolver-9wz2q" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572429 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-conf-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572466 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-host-var-lib-cni-bin\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572472 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-daemon-config\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572493 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-cnibin\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572539 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-os-release\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-socket-dir-parent\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572606 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-os-release\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9174d88f-f7f0-4b0c-942f-39d027b97e34-multus-cni-dir\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572692 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9174d88f-f7f0-4b0c-942f-39d027b97e34-cni-binary-copy\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572845 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572893 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-cni-binary-copy\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.572936 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-tuning-conf-dir\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.583943 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.592624 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b24hf\" (UniqueName: \"kubernetes.io/projected/9174d88f-f7f0-4b0c-942f-39d027b97e34-kube-api-access-b24hf\") pod \"multus-2wcn9\" (UID: \"9174d88f-f7f0-4b0c-942f-39d027b97e34\") " pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.594628 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzbc2\" (UniqueName: \"kubernetes.io/projected/5625d33c-6de7-4413-b9a5-e3ec2df2ef98-kube-api-access-wzbc2\") pod \"multus-additional-cni-plugins-wmhrj\" (UID: \"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\") " pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.597199 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtwvp\" (UniqueName: \"kubernetes.io/projected/af42f462-12a2-4488-b408-8753ac62bcf4-kube-api-access-gtwvp\") pod \"node-resolver-9wz2q\" (UID: \"af42f462-12a2-4488-b408-8753ac62bcf4\") " pod="openshift-dns/node-resolver-9wz2q" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.603689 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.615662 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.627061 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.638942 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.656812 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.658246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.658285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.658297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.658316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.658328 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.669024 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9wz2q" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.669100 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.675460 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2wcn9" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.679850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.683742 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: W1203 21:42:09.694109 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9174d88f_f7f0_4b0c_942f_39d027b97e34.slice/crio-28d9e5e9a6be057fd0973d700f38c3b67381026fe7ce345af62b8fc986bc5df8 WatchSource:0}: Error finding container 28d9e5e9a6be057fd0973d700f38c3b67381026fe7ce345af62b8fc986bc5df8: Status 404 returned error can't find the container with id 28d9e5e9a6be057fd0973d700f38c3b67381026fe7ce345af62b8fc986bc5df8 Dec 03 21:42:09 crc kubenswrapper[4715]: W1203 21:42:09.698420 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5625d33c_6de7_4413_b9a5_e3ec2df2ef98.slice/crio-48512ebf131c1204ab09ae5d17507e76582328bc2e6b9e65a61e08053943d5f4 WatchSource:0}: Error finding container 48512ebf131c1204ab09ae5d17507e76582328bc2e6b9e65a61e08053943d5f4: Status 404 returned error can't find the container with id 48512ebf131c1204ab09ae5d17507e76582328bc2e6b9e65a61e08053943d5f4 Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.706583 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.718924 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.729263 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.735869 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-ctls2"] Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.736191 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.743060 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.743916 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.744080 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8j8wk"] Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.744227 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.744477 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.744685 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.747087 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.749292 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.749738 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.749750 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.749749 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.749749 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.749945 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.749967 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.753108 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.766305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.766350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.766360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.766376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.766387 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.767614 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.783406 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.794254 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" event={"ID":"5625d33c-6de7-4413-b9a5-e3ec2df2ef98","Type":"ContainerStarted","Data":"48512ebf131c1204ab09ae5d17507e76582328bc2e6b9e65a61e08053943d5f4"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.796361 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2wcn9" event={"ID":"9174d88f-f7f0-4b0c-942f-39d027b97e34","Type":"ContainerStarted","Data":"28d9e5e9a6be057fd0973d700f38c3b67381026fe7ce345af62b8fc986bc5df8"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.797845 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9wz2q" event={"ID":"af42f462-12a2-4488-b408-8753ac62bcf4","Type":"ContainerStarted","Data":"50445f8cbcf1da46042655c47b0987bb5479daddadcec5abb416f296553c82a6"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.802857 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.815259 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.832896 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.847219 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.864616 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.873194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.873235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.873245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.873263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.873276 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878004 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-systemd-units\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878033 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-var-lib-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878053 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-env-overrides\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878082 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-script-lib\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878107 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-bin\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878121 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-node-log\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878135 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-etc-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-systemd\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878174 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba9d6059-9d98-4816-8fad-9e430e516d70-ovn-node-metrics-cert\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878189 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0ca0ce89-c3ec-421f-9954-c594760b16a9-rootfs\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878213 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-netns\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878229 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ca0ce89-c3ec-421f-9954-c594760b16a9-mcd-auth-proxy-config\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878245 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k26gp\" (UniqueName: \"kubernetes.io/projected/0ca0ce89-c3ec-421f-9954-c594760b16a9-kube-api-access-k26gp\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878261 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-ovn-kubernetes\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878290 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-kubelet\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878307 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-ovn\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878321 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-slash\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878333 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-log-socket\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878348 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ca0ce89-c3ec-421f-9954-c594760b16a9-proxy-tls\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878365 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878382 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpm4j\" (UniqueName: \"kubernetes.io/projected/ba9d6059-9d98-4816-8fad-9e430e516d70-kube-api-access-xpm4j\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878397 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878412 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-netd\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.878427 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-config\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.883233 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.899432 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.918732 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.932491 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.955261 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.976175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.976220 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.976237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.976258 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.976286 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:09Z","lastTransitionTime":"2025-12-03T21:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.978306 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.978863 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-bin\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.978916 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-node-log\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.978948 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-etc-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.978980 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-systemd\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979012 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba9d6059-9d98-4816-8fad-9e430e516d70-ovn-node-metrics-cert\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979043 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0ca0ce89-c3ec-421f-9954-c594760b16a9-rootfs\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979089 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-netns\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979122 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ca0ce89-c3ec-421f-9954-c594760b16a9-mcd-auth-proxy-config\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979158 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k26gp\" (UniqueName: \"kubernetes.io/projected/0ca0ce89-c3ec-421f-9954-c594760b16a9-kube-api-access-k26gp\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979191 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-ovn-kubernetes\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979238 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-kubelet\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979259 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-ovn\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979279 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-slash\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-log-socket\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ca0ce89-c3ec-421f-9954-c594760b16a9-proxy-tls\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979345 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979366 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpm4j\" (UniqueName: \"kubernetes.io/projected/ba9d6059-9d98-4816-8fad-9e430e516d70-kube-api-access-xpm4j\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979387 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979405 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-netd\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979424 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-config\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979455 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-systemd-units\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979477 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-var-lib-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979522 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-env-overrides\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.979543 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-script-lib\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.980310 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-script-lib\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.980378 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0ca0ce89-c3ec-421f-9954-c594760b16a9-rootfs\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.980414 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-netns\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981033 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0ca0ce89-c3ec-421f-9954-c594760b16a9-mcd-auth-proxy-config\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981306 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-ovn-kubernetes\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981349 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-kubelet\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981380 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-ovn\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981411 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-slash\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981452 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-log-socket\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981728 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-bin\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981864 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-node-log\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981902 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-etc-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.981954 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-systemd\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.982015 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-netd\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.982054 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.982544 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.982567 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-systemd-units\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.982625 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-var-lib-openvswitch\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.982931 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-env-overrides\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.983473 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-config\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.983968 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0ca0ce89-c3ec-421f-9954-c594760b16a9-proxy-tls\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.984752 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba9d6059-9d98-4816-8fad-9e430e516d70-ovn-node-metrics-cert\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.994227 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:09Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.996722 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k26gp\" (UniqueName: \"kubernetes.io/projected/0ca0ce89-c3ec-421f-9954-c594760b16a9-kube-api-access-k26gp\") pod \"machine-config-daemon-ctls2\" (UID: \"0ca0ce89-c3ec-421f-9954-c594760b16a9\") " pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:09 crc kubenswrapper[4715]: I1203 21:42:09.996815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpm4j\" (UniqueName: \"kubernetes.io/projected/ba9d6059-9d98-4816-8fad-9e430e516d70-kube-api-access-xpm4j\") pod \"ovnkube-node-8j8wk\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.057948 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.077498 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.079055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.079084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.079093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.079107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.079158 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: W1203 21:42:10.131558 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podba9d6059_9d98_4816_8fad_9e430e516d70.slice/crio-abb624dad13fd84bb42649176ed3e7cc27671d5a70a124f57d0baf90db00b41f WatchSource:0}: Error finding container abb624dad13fd84bb42649176ed3e7cc27671d5a70a124f57d0baf90db00b41f: Status 404 returned error can't find the container with id abb624dad13fd84bb42649176ed3e7cc27671d5a70a124f57d0baf90db00b41f Dec 03 21:42:10 crc kubenswrapper[4715]: W1203 21:42:10.133287 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ca0ce89_c3ec_421f_9954_c594760b16a9.slice/crio-1c8122e41fd9f677c8399a144dff5524ee8988f8dc5686b1c4ef028685de0fc9 WatchSource:0}: Error finding container 1c8122e41fd9f677c8399a144dff5524ee8988f8dc5686b1c4ef028685de0fc9: Status 404 returned error can't find the container with id 1c8122e41fd9f677c8399a144dff5524ee8988f8dc5686b1c4ef028685de0fc9 Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.189090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.189138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.189156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.189176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.189190 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.291300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.291332 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.291340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.291353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.291362 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.390162 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.390363 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.390448 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:42:18.390403753 +0000 UTC m=+35.133114388 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.390558 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.390602 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.390635 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.390661 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.390633 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.390743 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:18.390716831 +0000 UTC m=+35.133427466 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.390758 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.391017 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:18.390979699 +0000 UTC m=+35.133690504 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.390788 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.391127 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:18.391110482 +0000 UTC m=+35.133821287 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.391127 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.391240 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.391278 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.391296 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.391359 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:18.391344758 +0000 UTC m=+35.134055563 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.393965 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.394015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.394030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.394056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.394075 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.497131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.497631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.497648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.497671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.497687 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.600660 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.600699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.600713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.600730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.600744 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.633780 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.634046 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.634167 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.634262 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.634340 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:10 crc kubenswrapper[4715]: E1203 21:42:10.634422 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.703359 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.703402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.703414 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.703436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.703452 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.803177 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7" exitCode=0 Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.803283 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.803333 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"abb624dad13fd84bb42649176ed3e7cc27671d5a70a124f57d0baf90db00b41f"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.807935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.808000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.808020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.808051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.808070 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.808368 4715 generic.go:334] "Generic (PLEG): container finished" podID="5625d33c-6de7-4413-b9a5-e3ec2df2ef98" containerID="460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5" exitCode=0 Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.808487 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" event={"ID":"5625d33c-6de7-4413-b9a5-e3ec2df2ef98","Type":"ContainerDied","Data":"460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.812742 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2wcn9" event={"ID":"9174d88f-f7f0-4b0c-942f-39d027b97e34","Type":"ContainerStarted","Data":"0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.824106 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.825850 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9wz2q" event={"ID":"af42f462-12a2-4488-b408-8753ac62bcf4","Type":"ContainerStarted","Data":"72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.830422 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.830492 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.830543 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"1c8122e41fd9f677c8399a144dff5524ee8988f8dc5686b1c4ef028685de0fc9"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.850241 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.864406 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.881670 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.900829 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.917666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.917718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.917732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.917750 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.917763 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:10Z","lastTransitionTime":"2025-12-03T21:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.931947 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.946871 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.966895 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.980771 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:10 crc kubenswrapper[4715]: I1203 21:42:10.994220 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:10Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.009310 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.021481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.021570 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.021583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.021602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.021615 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.025054 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.037915 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.052490 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.071004 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.085093 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.118469 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.124313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.124342 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.124356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.124374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.124384 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.130086 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.153369 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.171107 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.188306 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.206420 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.220185 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.232975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.233014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.233026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.233044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.233056 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.238537 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.258839 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.280618 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.295236 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.311221 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.339855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.339912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.339929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.339951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.339966 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.442896 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.442954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.442970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.442993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.443008 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.545208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.545277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.545295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.545324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.545346 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.648350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.648413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.648431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.648458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.648477 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.754043 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.754115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.754133 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.754162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.754183 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.837068 4715 generic.go:334] "Generic (PLEG): container finished" podID="5625d33c-6de7-4413-b9a5-e3ec2df2ef98" containerID="9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b" exitCode=0 Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.837180 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" event={"ID":"5625d33c-6de7-4413-b9a5-e3ec2df2ef98","Type":"ContainerDied","Data":"9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.843184 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.843228 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.843253 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.843277 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.843295 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.843315 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.857749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.857810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.857829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.857856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.857877 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.863034 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.883148 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.903939 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.941106 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.961346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.961396 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.961410 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.961429 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.961446 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:11Z","lastTransitionTime":"2025-12-03T21:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.969051 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:11 crc kubenswrapper[4715]: I1203 21:42:11.989076 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:11Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.009920 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.026732 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.044882 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.063783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.063812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.063821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.063833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.063843 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.073016 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.083403 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.105989 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.118928 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.136786 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.166960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.166999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.167012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.167036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.167048 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.271091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.271175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.271198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.271231 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.271253 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.374783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.374855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.374876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.374908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.374928 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.477858 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.477915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.477933 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.477959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.477977 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.581217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.581278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.581296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.581321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.581340 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.633942 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.634021 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:12 crc kubenswrapper[4715]: E1203 21:42:12.634116 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.634137 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:12 crc kubenswrapper[4715]: E1203 21:42:12.634272 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:12 crc kubenswrapper[4715]: E1203 21:42:12.634373 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.684336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.684400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.684415 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.684436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.684448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.787257 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.787317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.787340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.787370 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.787391 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.852987 4715 generic.go:334] "Generic (PLEG): container finished" podID="5625d33c-6de7-4413-b9a5-e3ec2df2ef98" containerID="7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2" exitCode=0 Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.853060 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" event={"ID":"5625d33c-6de7-4413-b9a5-e3ec2df2ef98","Type":"ContainerDied","Data":"7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.883540 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.890614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.890692 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.890722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.890757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.890781 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.901588 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.927280 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.948079 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.963058 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.981467 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.992939 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.992997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.993010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.993030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.993044 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:12Z","lastTransitionTime":"2025-12-03T21:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:12 crc kubenswrapper[4715]: I1203 21:42:12.997360 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:12Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.014821 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.035731 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.046471 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.064800 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.082427 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.097010 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.097060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.097079 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.097105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.097126 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.101118 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.120091 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.199753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.199791 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.199809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.199830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.199844 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.303152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.303205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.303226 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.303254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.303277 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.406486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.406611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.406640 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.406673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.406695 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.510679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.510751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.510767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.510797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.510819 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.586734 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7sg46"] Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.587303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.590466 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.590735 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.591181 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.591232 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.616073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.616135 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.616153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.616179 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.616199 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.629203 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.647675 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.680607 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.700691 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.719614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.719672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.719695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.719725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.719743 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.720318 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.731150 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7fc36a21-545b-4055-92ae-6149beb86b89-serviceca\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.731401 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fc36a21-545b-4055-92ae-6149beb86b89-host\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.731601 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgcb9\" (UniqueName: \"kubernetes.io/projected/7fc36a21-545b-4055-92ae-6149beb86b89-kube-api-access-zgcb9\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.734588 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.757596 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.774369 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.792873 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.815882 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.822073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.822129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.822146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.822170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.822189 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.832477 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgcb9\" (UniqueName: \"kubernetes.io/projected/7fc36a21-545b-4055-92ae-6149beb86b89-kube-api-access-zgcb9\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.833067 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7fc36a21-545b-4055-92ae-6149beb86b89-serviceca\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.834700 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fc36a21-545b-4055-92ae-6149beb86b89-host\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.834844 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fc36a21-545b-4055-92ae-6149beb86b89-host\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.835007 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/7fc36a21-545b-4055-92ae-6149beb86b89-serviceca\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.843951 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.858767 4715 generic.go:334] "Generic (PLEG): container finished" podID="5625d33c-6de7-4413-b9a5-e3ec2df2ef98" containerID="db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7" exitCode=0 Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.858855 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" event={"ID":"5625d33c-6de7-4413-b9a5-e3ec2df2ef98","Type":"ContainerDied","Data":"db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.863073 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgcb9\" (UniqueName: \"kubernetes.io/projected/7fc36a21-545b-4055-92ae-6149beb86b89-kube-api-access-zgcb9\") pod \"node-ca-7sg46\" (UID: \"7fc36a21-545b-4055-92ae-6149beb86b89\") " pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.864245 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.886966 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.910163 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.915327 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7sg46" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.923653 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.925286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.925326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.925345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.925371 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.925388 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:13Z","lastTransitionTime":"2025-12-03T21:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.940901 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.955862 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.976764 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:13 crc kubenswrapper[4715]: I1203 21:42:13.999381 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:13Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.014543 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.025457 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.029476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.029562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.029581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.029609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.029628 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.044637 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.059161 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.074803 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.091438 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.114842 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.131669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.131704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.131712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.131727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.131737 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.132168 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.160468 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.172212 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.199221 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.223455 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.233873 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.234022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.234092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.234163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.234236 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.236600 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.261281 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.278557 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.295050 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.307359 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.321299 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.335175 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.336777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.336844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.336863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.336893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.336915 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.347717 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.363483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.384340 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.401891 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.419292 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.436967 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.440053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.440136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.440155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.440183 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.440202 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.458522 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.542555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.542619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.542633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.542661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.542680 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.633996 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.634017 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.634017 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:14 crc kubenswrapper[4715]: E1203 21:42:14.634175 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:14 crc kubenswrapper[4715]: E1203 21:42:14.634306 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:14 crc kubenswrapper[4715]: E1203 21:42:14.634610 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.645914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.645950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.645964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.645986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.646000 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.768856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.768949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.768969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.768999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.769019 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.867701 4715 generic.go:334] "Generic (PLEG): container finished" podID="5625d33c-6de7-4413-b9a5-e3ec2df2ef98" containerID="d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150" exitCode=0 Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.867807 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" event={"ID":"5625d33c-6de7-4413-b9a5-e3ec2df2ef98","Type":"ContainerDied","Data":"d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.873683 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7sg46" event={"ID":"7fc36a21-545b-4055-92ae-6149beb86b89","Type":"ContainerStarted","Data":"1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.873742 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7sg46" event={"ID":"7fc36a21-545b-4055-92ae-6149beb86b89","Type":"ContainerStarted","Data":"9eaf0e511ad2a37052dc359d218732444e7fc99be08fd0cdc7a0d1f6e7689dda"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.880159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.880225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.880247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.880277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.880297 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.889474 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.891312 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.913064 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.935412 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.963414 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.979857 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:14Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.983849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.983876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.983887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.983904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:14 crc kubenswrapper[4715]: I1203 21:42:14.983916 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:14Z","lastTransitionTime":"2025-12-03T21:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.011278 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.031310 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.054317 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.069470 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.087454 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.088138 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.088186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.088202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.088229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.088246 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.101438 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.117541 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.133789 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.144289 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.165310 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.185680 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.190018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.190050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.190058 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.190072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.190083 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.203804 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.227254 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.247601 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.268422 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.291958 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.293390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.293433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.293452 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.293479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.293498 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.327239 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.347290 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.363922 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.382120 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.396595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.396667 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.396693 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.396727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.396751 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.404048 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.428481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.445554 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.470382 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.495162 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.499712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.499782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.499803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.499846 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.499865 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.604393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.604465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.604488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.604552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.604580 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.634328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.634384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.634405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.634433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.634451 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: E1203 21:42:15.659405 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.665244 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.665307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.665327 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.665350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.665369 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: E1203 21:42:15.685257 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.690650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.690701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.690715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.690758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.690773 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: E1203 21:42:15.711936 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.717198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.717248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.717266 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.717289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.717305 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: E1203 21:42:15.737313 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.742561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.742638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.742657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.742685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.742704 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: E1203 21:42:15.763157 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: E1203 21:42:15.763413 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.766388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.766484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.766557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.766581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.766597 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.869863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.869932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.869950 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.869976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.869993 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.898993 4715 generic.go:334] "Generic (PLEG): container finished" podID="5625d33c-6de7-4413-b9a5-e3ec2df2ef98" containerID="b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371" exitCode=0 Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.899059 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" event={"ID":"5625d33c-6de7-4413-b9a5-e3ec2df2ef98","Type":"ContainerDied","Data":"b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371"} Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.921954 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.943266 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.961113 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.986123 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:15Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.995686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.995730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.995748 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.995773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:15 crc kubenswrapper[4715]: I1203 21:42:15.995790 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:15Z","lastTransitionTime":"2025-12-03T21:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.065840 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.079620 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.095096 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.100658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.100713 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.100733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.100760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.100781 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.113335 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.127437 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.140751 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.154618 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.166277 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.183454 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.194032 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.203288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.203315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.203325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.203338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.203349 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.213010 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.306223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.306284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.306301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.306325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.306342 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.409821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.409900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.409918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.409942 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.409961 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.513002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.513065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.513083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.513110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.513129 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.616487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.616592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.616614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.616645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.616665 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.633839 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.633921 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.633844 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:16 crc kubenswrapper[4715]: E1203 21:42:16.634085 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:16 crc kubenswrapper[4715]: E1203 21:42:16.633933 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:16 crc kubenswrapper[4715]: E1203 21:42:16.634241 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.720819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.720883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.720902 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.720928 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.720948 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.824375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.824443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.824462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.824489 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.824534 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.908105 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" event={"ID":"5625d33c-6de7-4413-b9a5-e3ec2df2ef98","Type":"ContainerStarted","Data":"379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.917719 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.920088 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.920161 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.927057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.927090 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.927103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.927118 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.927130 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:16Z","lastTransitionTime":"2025-12-03T21:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.927753 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.954983 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.962127 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.963948 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.976969 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:16 crc kubenswrapper[4715]: I1203 21:42:16.997604 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:16Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.021201 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.031574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.031666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.031694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.031736 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.031765 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.044351 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.077419 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.100734 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.124201 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.139176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.139242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.139278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.139557 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.139596 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.164139 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.185052 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.221587 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.241159 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.242690 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.242723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.242735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.242752 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.242765 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.260669 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.285750 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.308458 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.328433 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.346003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.346065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.346088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.346116 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.346139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.349609 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.370624 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.387775 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.420102 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.438222 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.449604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.449739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.449765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.449796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.449816 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.473266 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.494949 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.512137 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.525789 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.545936 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.553411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.553461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.553478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.553534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.553552 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.569244 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.586814 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.610746 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:17Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.655632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.655691 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.655709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.655735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.655753 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.758705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.758749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.758761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.758778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.758792 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.862291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.862358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.862376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.862399 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.862415 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.922003 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.965295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.965360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.965378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.965404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:17 crc kubenswrapper[4715]: I1203 21:42:17.965423 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:17Z","lastTransitionTime":"2025-12-03T21:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.068774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.068859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.068876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.068901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.068919 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.173493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.173563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.173575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.173592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.173603 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.283018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.283112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.283141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.283180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.283214 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.407817 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.407944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408031 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:42:34.407986553 +0000 UTC m=+51.150697148 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408075 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408096 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408111 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408181 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:34.408160577 +0000 UTC m=+51.150871302 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.408182 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.408241 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.408273 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408350 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408368 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408375 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408383 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408422 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:34.408413734 +0000 UTC m=+51.151124329 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408448 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:34.408429214 +0000 UTC m=+51.151139819 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408606 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.408762 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:34.408715142 +0000 UTC m=+51.151425977 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.409714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.409754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.409767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.409787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.409801 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.511762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.511789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.511797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.511810 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.511818 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.614280 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.614309 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.614317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.614330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.614339 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.634130 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.634223 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.634493 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.634561 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.634600 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:18 crc kubenswrapper[4715]: E1203 21:42:18.634640 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.717854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.717898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.717916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.717940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.717956 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.821454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.821510 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.821606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.821635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.821676 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.924917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.924976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.924995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.925020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.925036 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:18Z","lastTransitionTime":"2025-12-03T21:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:18 crc kubenswrapper[4715]: I1203 21:42:18.926607 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.027856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.027917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.027959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.027982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.027999 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.131914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.131953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.131968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.131989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.132004 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.235705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.235773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.235797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.235829 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.235851 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.339639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.339694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.339712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.339733 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.339749 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.443754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.443823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.443843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.443871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.443890 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.546961 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.547015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.547033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.547063 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.547080 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.650114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.650253 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.650286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.650317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.650339 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.753136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.753214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.753243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.753275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.753300 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.857052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.857106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.857117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.857142 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.857156 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.932879 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/0.log" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.937678 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa" exitCode=1 Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.937754 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.938962 4715 scope.go:117] "RemoveContainer" containerID="b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.960052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.960110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.960129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.960158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.960180 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:19Z","lastTransitionTime":"2025-12-03T21:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.971548 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:19Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:19 crc kubenswrapper[4715]: I1203 21:42:19.989941 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:19Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.023069 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.043979 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.060889 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.067973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.068019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.068034 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.068055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.068073 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.076497 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.097952 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.115004 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.128342 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.137053 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.148654 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.159064 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.171584 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.171613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.171624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.171641 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.171653 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.175381 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.193909 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.211909 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.274096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.274149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.274169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.274196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.274214 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.291122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.306302 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.321571 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.332413 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.343224 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.358391 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.370340 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.377361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.377408 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.377420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.377436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.377448 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.388374 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.402778 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.416147 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.428077 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.441644 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.453996 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.470462 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.479459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.479483 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.479494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.479528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.479541 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.482281 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.504228 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.581632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.581697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.581715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.581740 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.581760 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.633851 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:20 crc kubenswrapper[4715]: E1203 21:42:20.633976 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.634326 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:20 crc kubenswrapper[4715]: E1203 21:42:20.634373 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.634406 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:20 crc kubenswrapper[4715]: E1203 21:42:20.634448 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.684175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.684200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.684207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.684219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.684228 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.787053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.787088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.787100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.787115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.787126 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.889602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.889633 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.889644 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.889657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.889666 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.943709 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/0.log" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.947233 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.947367 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.969704 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.993550 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.993616 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.993642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.993673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.993696 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:20Z","lastTransitionTime":"2025-12-03T21:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:20 crc kubenswrapper[4715]: I1203 21:42:20.999734 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:20Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.018722 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.048914 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.061693 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.089585 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.097331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.097395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.097422 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.097455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.097479 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.105904 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.134176 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.149133 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.162369 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.174011 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.185230 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.199872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.199915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.199927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.199979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.199991 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.202285 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.218909 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.232650 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:21Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.301964 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.302100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.302112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.302128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.302139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.409298 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.409582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.409665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.409744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.409842 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.512890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.512951 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.512968 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.512996 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.513016 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.616345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.616411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.616430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.616456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.616474 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.719482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.719547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.719561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.719577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.719590 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.822551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.822618 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.822632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.822673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.822684 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.926600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.926714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.926742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.926794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:21 crc kubenswrapper[4715]: I1203 21:42:21.926829 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:21Z","lastTransitionTime":"2025-12-03T21:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.029807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.029876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.029897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.029925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.029944 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.134091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.134174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.134196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.134224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.134244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.237482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.237592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.237609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.237637 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.237656 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.341482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.341596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.341614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.341638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.341653 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.378793 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8"] Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.379493 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.382288 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.384720 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.417104 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.432093 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.444582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.444634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.444652 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.444677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.444694 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.454969 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.456867 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bljbm\" (UniqueName: \"kubernetes.io/projected/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-kube-api-access-bljbm\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.456928 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.456995 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.457046 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.469471 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.481012 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.493298 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.502963 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.518006 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.532113 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.545345 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.547091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.547190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.547210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.547233 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.547248 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.558284 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.558355 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.558399 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bljbm\" (UniqueName: \"kubernetes.io/projected/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-kube-api-access-bljbm\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.558430 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.559968 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.560364 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-env-overrides\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.567850 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.569609 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.576461 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bljbm\" (UniqueName: \"kubernetes.io/projected/ea6cedd1-c3cc-4408-b6f6-a58fcac998a9-kube-api-access-bljbm\") pod \"ovnkube-control-plane-749d76644c-47tx8\" (UID: \"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.583764 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.598434 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.612739 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.628185 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.633774 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.633799 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.633779 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:22 crc kubenswrapper[4715]: E1203 21:42:22.633923 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:22 crc kubenswrapper[4715]: E1203 21:42:22.634068 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:22 crc kubenswrapper[4715]: E1203 21:42:22.634190 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.642642 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.649495 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.649596 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.649610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.649630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.649644 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.697344 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" Dec 03 21:42:22 crc kubenswrapper[4715]: W1203 21:42:22.713481 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea6cedd1_c3cc_4408_b6f6_a58fcac998a9.slice/crio-4a8a358c94f9d74e59f14a60026fb76234baa16f1d46ac88ec44b47ce5ace31c WatchSource:0}: Error finding container 4a8a358c94f9d74e59f14a60026fb76234baa16f1d46ac88ec44b47ce5ace31c: Status 404 returned error can't find the container with id 4a8a358c94f9d74e59f14a60026fb76234baa16f1d46ac88ec44b47ce5ace31c Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.759997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.760031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.760040 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.760054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.760063 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.861828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.861874 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.861889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.861910 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.861925 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.956230 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" event={"ID":"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9","Type":"ContainerStarted","Data":"4a8a358c94f9d74e59f14a60026fb76234baa16f1d46ac88ec44b47ce5ace31c"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.957997 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/1.log" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.958643 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/0.log" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.961462 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6" exitCode=1 Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.961490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.961549 4715 scope.go:117] "RemoveContainer" containerID="b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.962243 4715 scope.go:117] "RemoveContainer" containerID="65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6" Dec 03 21:42:22 crc kubenswrapper[4715]: E1203 21:42:22.962502 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.963777 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.963803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.963811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.963824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.963833 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:22Z","lastTransitionTime":"2025-12-03T21:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.976483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:22 crc kubenswrapper[4715]: I1203 21:42:22.991583 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:22Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.005038 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.050314 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.069769 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.069818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.069830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.069850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.069862 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.078974 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.090991 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.107924 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.122524 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.138664 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.148199 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.161713 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.172454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.172479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.172487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.172514 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.172523 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.173933 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.188782 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.210083 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.227754 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.243665 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.274821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.274840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.274849 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.274861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.274871 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.378088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.378163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.378187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.378221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.378248 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.480855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.481069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.481078 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.481093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.481103 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.510734 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-hbsmx"] Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.511346 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:23 crc kubenswrapper[4715]: E1203 21:42:23.511443 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.526740 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.561702 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.565804 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg2g5\" (UniqueName: \"kubernetes.io/projected/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-kube-api-access-dg2g5\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.565882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.584490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.584586 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.584605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.584632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.584649 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.619459 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.645060 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.659154 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.667291 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg2g5\" (UniqueName: \"kubernetes.io/projected/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-kube-api-access-dg2g5\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.667383 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:23 crc kubenswrapper[4715]: E1203 21:42:23.667637 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:23 crc kubenswrapper[4715]: E1203 21:42:23.667718 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs podName:34ccf04b-c534-43bb-94ee-fc2e7cef68c6 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:24.167695409 +0000 UTC m=+40.910406034 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs") pod "network-metrics-daemon-hbsmx" (UID: "34ccf04b-c534-43bb-94ee-fc2e7cef68c6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.674227 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.690484 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.690550 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.690560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.690576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.690591 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.692710 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg2g5\" (UniqueName: \"kubernetes.io/projected/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-kube-api-access-dg2g5\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.705615 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.721474 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.738916 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.765556 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.781104 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.793669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.793729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.793746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.793771 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.793788 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.797483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.814677 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.829783 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.846847 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.867265 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.888146 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.897122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.897159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.897167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.897182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.897193 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.904465 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.924292 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.938178 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.957480 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.966378 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/1.log" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.970545 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" event={"ID":"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9","Type":"ContainerStarted","Data":"ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.970578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" event={"ID":"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9","Type":"ContainerStarted","Data":"6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b"} Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.983390 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:23Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.999695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.999744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.999760 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:23 crc kubenswrapper[4715]: I1203 21:42:23.999782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:23.999798 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:23Z","lastTransitionTime":"2025-12-03T21:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.002807 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.021243 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.041060 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.058691 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.082021 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.100607 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.102339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.102393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.102410 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.102436 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.102455 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.135645 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.155175 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.171272 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:24 crc kubenswrapper[4715]: E1203 21:42:24.171491 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:24 crc kubenswrapper[4715]: E1203 21:42:24.171652 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs podName:34ccf04b-c534-43bb-94ee-fc2e7cef68c6 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:25.171622261 +0000 UTC m=+41.914332886 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs") pod "network-metrics-daemon-hbsmx" (UID: "34ccf04b-c534-43bb-94ee-fc2e7cef68c6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.172406 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.191559 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.204792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.204853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.204872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.204897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.204914 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.210622 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.224940 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.242633 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.257796 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.271707 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.286486 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.307493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.307563 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.307582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.307606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.307626 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.308221 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.327053 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.347612 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.363003 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.377582 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.389357 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.409095 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.409948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.410006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.410019 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.410036 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.410047 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.421961 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.436113 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.448694 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.476862 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.491172 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.512766 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:24Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.513154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.513211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.513232 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.513259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.513278 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.616214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.616270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.616287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.616311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.616328 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.633823 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.633866 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:24 crc kubenswrapper[4715]: E1203 21:42:24.633998 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.634141 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:24 crc kubenswrapper[4715]: E1203 21:42:24.634294 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.634450 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:24 crc kubenswrapper[4715]: E1203 21:42:24.634593 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:24 crc kubenswrapper[4715]: E1203 21:42:24.634720 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.718318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.718379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.718425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.718461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.718485 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.821541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.821582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.821592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.821607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.821617 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.924732 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.924796 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.924814 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.924841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:24 crc kubenswrapper[4715]: I1203 21:42:24.924859 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:24Z","lastTransitionTime":"2025-12-03T21:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.027542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.027603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.027621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.027649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.027668 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.129866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.129922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.129940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.129963 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.129981 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.179353 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:25 crc kubenswrapper[4715]: E1203 21:42:25.179592 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:25 crc kubenswrapper[4715]: E1203 21:42:25.179697 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs podName:34ccf04b-c534-43bb-94ee-fc2e7cef68c6 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:27.17967632 +0000 UTC m=+43.922386915 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs") pod "network-metrics-daemon-hbsmx" (UID: "34ccf04b-c534-43bb-94ee-fc2e7cef68c6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.232761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.232825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.232844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.232871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.232894 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.335456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.335549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.335568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.335593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.335612 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.438915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.438972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.438988 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.439011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.439029 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.542699 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.542763 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.542787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.542821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.542844 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.646050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.646101 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.646117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.646141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.646157 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.749256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.749307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.749322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.749344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.749360 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.852768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.852819 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.852836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.852861 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.852878 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.929828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.929897 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.929918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.929940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.929958 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: E1203 21:42:25.952051 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:25Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.957131 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.957196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.957215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.957241 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.957259 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:25 crc kubenswrapper[4715]: E1203 21:42:25.978357 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:25Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.983004 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.983056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.983072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.983099 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:25 crc kubenswrapper[4715]: I1203 21:42:25.983120 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:25Z","lastTransitionTime":"2025-12-03T21:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: E1203 21:42:26.004881 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:26Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.011117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.011173 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.011191 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.011216 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.011310 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: E1203 21:42:26.031951 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:26Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.036971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.037034 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.037052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.037077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.037096 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: E1203 21:42:26.058272 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:26Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:26 crc kubenswrapper[4715]: E1203 21:42:26.058600 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.061476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.061554 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.061575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.061601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.061618 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.164336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.164385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.164402 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.164426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.164443 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.266980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.267029 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.267042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.267059 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.267072 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.369558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.369594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.369602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.369614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.369622 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.471825 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.471962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.471992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.472024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.472045 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.574842 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.574892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.574904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.574920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.574931 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.634197 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.634280 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.634329 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:26 crc kubenswrapper[4715]: E1203 21:42:26.634599 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.634652 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:26 crc kubenswrapper[4715]: E1203 21:42:26.634743 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:26 crc kubenswrapper[4715]: E1203 21:42:26.634862 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:26 crc kubenswrapper[4715]: E1203 21:42:26.635126 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.678054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.678115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.678133 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.678158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.678178 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.781972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.782039 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.782056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.782083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.782099 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.885655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.885719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.885737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.885765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.885786 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.989293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.989348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.989366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.989391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:26 crc kubenswrapper[4715]: I1203 21:42:26.989409 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:26Z","lastTransitionTime":"2025-12-03T21:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.092613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.093021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.093103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.093169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.093228 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.197350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.197419 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.197439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.197467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.197488 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.203012 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:27 crc kubenswrapper[4715]: E1203 21:42:27.203291 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:27 crc kubenswrapper[4715]: E1203 21:42:27.203439 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs podName:34ccf04b-c534-43bb-94ee-fc2e7cef68c6 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:31.20339706 +0000 UTC m=+47.946107695 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs") pod "network-metrics-daemon-hbsmx" (UID: "34ccf04b-c534-43bb-94ee-fc2e7cef68c6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.301383 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.301498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.301555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.301585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.301606 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.404074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.404235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.404256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.404329 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.404351 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.507447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.507552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.507573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.507598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.507618 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.611137 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.611196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.611212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.611238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.611255 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.713828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.713884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.713902 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.713925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.713944 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.816900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.816962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.816975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.816993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.817008 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.920325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.920392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.920409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.920431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:27 crc kubenswrapper[4715]: I1203 21:42:27.920445 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:27Z","lastTransitionTime":"2025-12-03T21:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.024702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.024774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.024800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.024831 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.024852 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.127715 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.127772 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.127790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.127816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.127835 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.230321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.230375 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.230392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.230417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.230441 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.333209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.333300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.333325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.333357 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.333388 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.436880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.437351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.437575 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.437784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.437940 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.541475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.541537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.541547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.541562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.541571 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.633880 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.633930 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:28 crc kubenswrapper[4715]: E1203 21:42:28.633983 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:28 crc kubenswrapper[4715]: E1203 21:42:28.634100 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.634447 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.634600 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:28 crc kubenswrapper[4715]: E1203 21:42:28.634938 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:28 crc kubenswrapper[4715]: E1203 21:42:28.635114 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.644208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.644290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.644315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.644341 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.644364 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.746907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.746975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.746997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.747031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.747055 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.850130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.850183 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.850199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.850225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.850243 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.953441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.953542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.953565 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.953594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:28 crc kubenswrapper[4715]: I1203 21:42:28.953611 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:28Z","lastTransitionTime":"2025-12-03T21:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.057917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.057990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.058013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.058042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.058064 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.161077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.161343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.161458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.161584 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.161667 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.265328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.265645 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.265734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.265866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.265965 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.368729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.369106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.369308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.369488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.369804 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.473343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.473391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.473407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.473432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.473449 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.576827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.576876 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.576893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.576916 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.576933 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.680447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.680567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.680603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.680634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.680657 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.783947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.784088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.784108 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.784130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.784147 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.887162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.887223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.887248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.887276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.887297 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.990856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.990932 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.990959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.990988 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:29 crc kubenswrapper[4715]: I1203 21:42:29.991011 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:29Z","lastTransitionTime":"2025-12-03T21:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.094219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.094286 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.094307 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.094338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.094360 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.197239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.197295 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.197315 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.197346 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.197369 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.300770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.300818 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.300836 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.300860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.300877 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.404156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.404219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.404237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.404263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.404281 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.491633 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.503363 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.507666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.507735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.507757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.507788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.507811 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.512813 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.543964 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b99dbec09b348bd3f4278b942932233bde7f23cf57524cf5c4d02f6131984ffa\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:19Z\\\",\\\"message\\\":\\\"42:19.086439 6032 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 21:42:19.086600 6032 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.086828 6032 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 21:42:19.087068 6032 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087159 6032 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087259 6032 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087515 6032 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.087946 6032 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:19.088006 6032 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.561853 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.578991 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.611954 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.613179 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.613264 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.613333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.613360 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.614386 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.631589 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.633850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:30 crc kubenswrapper[4715]: E1203 21:42:30.633985 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.634064 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:30 crc kubenswrapper[4715]: E1203 21:42:30.634158 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.634224 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:30 crc kubenswrapper[4715]: E1203 21:42:30.634297 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.634351 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:30 crc kubenswrapper[4715]: E1203 21:42:30.634426 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.650483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.670401 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.690873 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.711653 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.717169 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.717463 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.717712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.717880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.718026 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.727749 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.756804 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.777970 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.800818 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.820828 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.820977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.821024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.821041 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.821065 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.821083 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.836969 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.857931 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:30Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.924391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.924448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.924464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.924486 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:30 crc kubenswrapper[4715]: I1203 21:42:30.924533 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:30Z","lastTransitionTime":"2025-12-03T21:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.028579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.028638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.028655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.028678 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.028695 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.131321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.131387 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.131405 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.131430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.131447 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.233625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.233892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.233979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.234080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.234170 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.247452 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:31 crc kubenswrapper[4715]: E1203 21:42:31.247736 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:31 crc kubenswrapper[4715]: E1203 21:42:31.247830 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs podName:34ccf04b-c534-43bb-94ee-fc2e7cef68c6 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:39.247807859 +0000 UTC m=+55.990518494 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs") pod "network-metrics-daemon-hbsmx" (UID: "34ccf04b-c534-43bb-94ee-fc2e7cef68c6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.336455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.336552 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.336571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.336597 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.336615 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.439921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.439985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.440001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.440026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.440047 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.543144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.543198 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.543214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.543239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.543256 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.646183 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.646249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.646268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.646293 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.646311 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.749710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.749863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.749886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.749911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.749930 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.852464 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.852562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.852584 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.852614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.852634 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.955559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.955622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.955641 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.955672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:31 crc kubenswrapper[4715]: I1203 21:42:31.955693 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:31Z","lastTransitionTime":"2025-12-03T21:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.058946 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.059032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.059051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.059109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.059128 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.162923 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.162990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.163007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.163037 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.163058 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.265697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.265762 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.265856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.265885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.265907 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.368929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.369009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.369028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.369056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.369077 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.471743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.471809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.471828 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.471855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.471885 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.574687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.574749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.574767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.574793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.574813 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.633690 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.633762 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.633798 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:32 crc kubenswrapper[4715]: E1203 21:42:32.633977 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.634061 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.634064 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:32 crc kubenswrapper[4715]: E1203 21:42:32.634179 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:32 crc kubenswrapper[4715]: E1203 21:42:32.634561 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:32 crc kubenswrapper[4715]: E1203 21:42:32.634710 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.635775 4715 scope.go:117] "RemoveContainer" containerID="65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.654895 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.675600 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.679441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.679527 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.679550 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.679576 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.679594 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.695289 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.716892 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.737398 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.758134 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.783212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.783283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.783301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.783354 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.783793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.783824 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.807303 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.833198 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.853016 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.876638 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.887123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.887165 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.887184 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.887204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.887222 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.892029 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.905891 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.927458 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.942816 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.961485 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.977125 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.989444 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:32Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.989900 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.989917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.989927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.989944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:32 crc kubenswrapper[4715]: I1203 21:42:32.989955 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:32Z","lastTransitionTime":"2025-12-03T21:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.093542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.093592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.093642 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.093670 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.093686 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.196613 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.196648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.196657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.196672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.196682 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.299337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.299365 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.299373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.299385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.299394 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.401543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.401590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.401602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.401620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.401634 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.504943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.504990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.505002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.505021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.505033 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.607957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.607997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.608006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.608023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.608034 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.661277 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.680601 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.710786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.710830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.710841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.710860 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.710932 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.733313 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.772483 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.787941 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.801086 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.813979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.814026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.814039 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.814061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.814075 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.820739 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.833764 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.850790 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.865207 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.877273 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.894195 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.910703 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.916204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.916276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.916289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.916314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.916329 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:33Z","lastTransitionTime":"2025-12-03T21:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.934293 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.960842 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:33 crc kubenswrapper[4715]: I1203 21:42:33.984297 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:33Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.007025 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.013085 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/1.log" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.017968 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.018790 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.018837 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.018856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.018882 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.018904 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.019854 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.028481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.050278 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.073295 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.091443 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.114401 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.121929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.121962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.121977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.121998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.122013 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.132087 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.153295 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.172240 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.192092 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.225695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.226397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.226577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.226755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.226891 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.232224 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.245404 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.265885 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.287771 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.304925 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.317846 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.332066 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.332112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.332129 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.332155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.332172 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.335378 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.356799 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.378687 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.394071 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:34Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.435976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.436029 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.436046 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.436070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.436088 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.484055 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.484151 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484200 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:43:06.484171225 +0000 UTC m=+83.226881860 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484267 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484284 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484296 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.484294 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.484342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.484389 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484490 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484604 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 21:43:06.484541386 +0000 UTC m=+83.227252021 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484666 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484687 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:43:06.484670959 +0000 UTC m=+83.227381594 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484688 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484717 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.484772 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 21:43:06.484757561 +0000 UTC m=+83.227468187 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.485581 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.485655 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:43:06.485640614 +0000 UTC m=+83.228351239 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.539049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.539110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.539130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.539163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.539186 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.633526 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.633603 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.633614 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.633659 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.633692 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.633772 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.633954 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:34 crc kubenswrapper[4715]: E1203 21:42:34.634066 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.641898 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.641927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.641937 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.641952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.641962 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.745104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.745151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.745163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.745180 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.745192 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.847724 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.847758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.847768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.847782 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.847791 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.950535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.950676 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.950695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.950721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:34 crc kubenswrapper[4715]: I1203 21:42:34.950738 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:34Z","lastTransitionTime":"2025-12-03T21:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.025713 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/2.log" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.026590 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/1.log" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.030752 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4" exitCode=1 Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.030812 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.030875 4715 scope.go:117] "RemoveContainer" containerID="65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.032180 4715 scope.go:117] "RemoveContainer" containerID="2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4" Dec 03 21:42:35 crc kubenswrapper[4715]: E1203 21:42:35.032566 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.052295 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.053192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.053240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.053255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.053276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.053292 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.069133 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.085123 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.102229 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.117086 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.138114 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.150326 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.155854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.155941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.155959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.156012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.156035 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.171616 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.186015 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.212608 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65441842d917df7011acec8b195fdbbe78b4155ebef51de9b177d4460ca10ec6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 21:42:21.096590 6161 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1203 21:42:21.096620 6161 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1203 21:42:21.096626 6161 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1203 21:42:21.096639 6161 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1203 21:42:21.096643 6161 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 21:42:21.096680 6161 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 21:42:21.096691 6161 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1203 21:42:21.096687 6161 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 21:42:21.096696 6161 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 21:42:21.096706 6161 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1203 21:42:21.096723 6161 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1203 21:42:21.096726 6161 handler.go:208] Removed *v1.Node event handler 7\\\\nI1203 21:42:21.096755 6161 factory.go:656] Stopping watch factory\\\\nI1203 21:42:21.096755 6161 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 21:42:21.096768 6161 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:20Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.227485 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.245717 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.259106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.259287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.259379 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.259477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.259641 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.261672 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.279332 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.302560 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.319932 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.336402 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.352608 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:35Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.362906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.363204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.363648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.363843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.364034 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.467330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.467697 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.467880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.468032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.468150 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.571049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.571098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.571113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.571134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.571149 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.673160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.673248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.673270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.673296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.673316 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.776006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.776337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.776487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.776737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.776886 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.879940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.879983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.879996 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.880013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.880024 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.983158 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.983300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.983313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.983330 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:35 crc kubenswrapper[4715]: I1203 21:42:35.983342 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:35Z","lastTransitionTime":"2025-12-03T21:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.035555 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/2.log" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.040837 4715 scope.go:117] "RemoveContainer" containerID="2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.041283 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.062974 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.083536 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.086220 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.086257 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.086272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.086314 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.086339 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.098746 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.116254 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.136894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.136978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.136999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.137031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.137054 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.138934 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.154959 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.155474 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.160001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.160033 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.160044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.160061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.160073 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.173847 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.215258 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.220406 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.220432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.220442 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.220458 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.220467 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.231311 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.237064 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.242466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.242492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.242604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.242657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.242703 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.249990 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.259147 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.263105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.263154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.263174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.263199 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.263216 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.269495 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.284830 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.284955 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.286954 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.287007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.287020 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.287055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.287071 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.287491 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.310411 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.333786 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.367821 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.386705 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.392190 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.392401 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.392543 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.392685 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.392792 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.422203 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.444375 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.461545 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:36Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.497024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.497085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.497104 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.497136 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.497158 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.601168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.601246 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.601269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.601303 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.601326 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.634682 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.634762 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.634751 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.635205 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.635387 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.635680 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.635833 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:36 crc kubenswrapper[4715]: E1203 21:42:36.635983 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.706409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.706481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.706528 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.706560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.706580 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.810390 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.810480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.810541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.810573 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.810593 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.914855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.914907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.914920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.914943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:36 crc kubenswrapper[4715]: I1203 21:42:36.914958 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:36Z","lastTransitionTime":"2025-12-03T21:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.019210 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.019274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.019289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.019313 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.019329 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.122461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.122535 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.122549 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.122572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.122588 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.226132 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.226191 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.226209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.226237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.226256 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.329608 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.329694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.329719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.329754 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.329780 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.433234 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.433304 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.433323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.433350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.433402 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.536915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.536977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.536992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.537011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.537024 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.640674 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.640745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.640759 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.640778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.640796 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.744491 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.744614 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.744627 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.744648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.744663 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.848213 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.848282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.848305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.848336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.848355 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.952222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.952275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.952284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.952300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:37 crc kubenswrapper[4715]: I1203 21:42:37.952310 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:37Z","lastTransitionTime":"2025-12-03T21:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.054766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.054835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.054852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.054904 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.054926 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.158191 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.158247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.158267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.158300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.158322 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.263240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.263291 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.263305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.263326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.263340 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.365925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.365979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.365992 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.366009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.366022 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.469480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.469605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.469623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.469650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.469669 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.578907 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.578980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.579002 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.579030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.579050 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.634166 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.634202 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.634331 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.634367 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:38 crc kubenswrapper[4715]: E1203 21:42:38.634618 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:38 crc kubenswrapper[4715]: E1203 21:42:38.634770 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:38 crc kubenswrapper[4715]: E1203 21:42:38.634876 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:38 crc kubenswrapper[4715]: E1203 21:42:38.635004 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.682263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.682317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.682328 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.682349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.682364 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.786725 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.786806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.786830 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.786865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.786922 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.892153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.892186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.892195 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.892207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.892217 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.995125 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.995207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.995229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.995259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:38 crc kubenswrapper[4715]: I1203 21:42:38.995278 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:38Z","lastTransitionTime":"2025-12-03T21:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.099238 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.099324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.099345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.099377 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.099398 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.202585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.202694 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.202714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.202780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.202802 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.252893 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:39 crc kubenswrapper[4715]: E1203 21:42:39.253237 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:39 crc kubenswrapper[4715]: E1203 21:42:39.253397 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs podName:34ccf04b-c534-43bb-94ee-fc2e7cef68c6 nodeName:}" failed. No retries permitted until 2025-12-03 21:42:55.253359429 +0000 UTC m=+71.996070064 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs") pod "network-metrics-daemon-hbsmx" (UID: "34ccf04b-c534-43bb-94ee-fc2e7cef68c6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.307696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.308308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.308333 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.308366 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.308390 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.412696 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.412821 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.412850 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.412887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.412912 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.515765 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.515893 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.515947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.515975 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.515995 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.620098 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.620192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.620218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.620256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.620282 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.723885 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.723962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.723980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.724007 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.724027 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.827863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.827969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.827990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.828016 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.828041 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.932339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.932409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.932426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.932452 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:39 crc kubenswrapper[4715]: I1203 21:42:39.932469 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:39Z","lastTransitionTime":"2025-12-03T21:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.036583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.036658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.036677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.036701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.036722 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.140023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.140086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.140105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.140134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.140153 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.245686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.245737 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.245755 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.245778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.245802 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.349321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.349378 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.349545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.349578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.349598 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.452986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.453050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.453062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.453075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.453086 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.556479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.556593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.556617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.556654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.556678 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.634115 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.634178 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:40 crc kubenswrapper[4715]: E1203 21:42:40.634322 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.634101 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.634377 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:40 crc kubenswrapper[4715]: E1203 21:42:40.634570 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:40 crc kubenswrapper[4715]: E1203 21:42:40.634691 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:40 crc kubenswrapper[4715]: E1203 21:42:40.634773 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.659606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.659664 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.659683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.659710 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.659733 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.770536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.770593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.770611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.770636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.770676 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.873498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.873590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.873609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.873631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.873648 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.977008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.977077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.977096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.977122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:40 crc kubenswrapper[4715]: I1203 21:42:40.977146 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:40Z","lastTransitionTime":"2025-12-03T21:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.079997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.080063 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.080080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.080107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.080121 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.183477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.183568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.183587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.183617 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.183635 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.287364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.287438 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.287456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.287478 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.287494 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.393115 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.393555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.393571 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.393595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.393610 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.497409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.497551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.497581 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.497659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.497683 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.601348 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.601434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.601454 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.601487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.601541 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.704164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.704196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.704204 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.704218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.704227 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.807747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.807822 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.807834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.807859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.807874 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.911537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.911598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.911621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.911646 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:41 crc kubenswrapper[4715]: I1203 21:42:41.911666 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:41Z","lastTransitionTime":"2025-12-03T21:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.015807 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.015870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.015890 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.015914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.015932 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.119707 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.119768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.119786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.119816 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.119836 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.223411 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.223481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.223540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.223689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.223723 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.327153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.327223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.327242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.327282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.327301 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.429857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.429917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.429930 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.429947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.429959 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.532677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.532735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.532757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.532788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.532809 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.633884 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:42 crc kubenswrapper[4715]: E1203 21:42:42.634064 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.634130 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.634253 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:42 crc kubenswrapper[4715]: E1203 21:42:42.634435 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.634565 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:42 crc kubenswrapper[4715]: E1203 21:42:42.634899 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:42 crc kubenswrapper[4715]: E1203 21:42:42.635159 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.636146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.636209 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.636265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.636301 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.636324 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.740205 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.740261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.740279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.740305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.740324 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.845096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.845155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.845167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.845192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.845208 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.955531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.955626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.955643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.955661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:42 crc kubenswrapper[4715]: I1203 21:42:42.955700 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:42Z","lastTransitionTime":"2025-12-03T21:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.059265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.059320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.059338 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.059361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.059378 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.163619 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.163700 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.163721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.163757 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.163783 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.268208 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.268300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.268388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.268482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.268541 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.371767 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.371856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.371883 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.371923 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.371947 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.474947 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.475001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.475018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.475042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.475061 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.578153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.578227 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.578249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.578282 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.578306 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.659869 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.681946 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.686023 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.686120 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.686143 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.686182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.686215 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.700754 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.766785 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.789088 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.789149 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.789167 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.789192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.789212 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.803352 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.823942 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.844334 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.869239 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.891652 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.894871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.894955 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.894973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.895001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.895020 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.911700 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.959707 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.980263 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:43Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.998969 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.999052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.999072 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.999100 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:43 crc kubenswrapper[4715]: I1203 21:42:43.999119 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:43Z","lastTransitionTime":"2025-12-03T21:42:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.015763 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:44Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.035488 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:44Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.056620 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:44Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.074126 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:44Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.090851 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:44Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.102605 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.102815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.102955 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.103092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.103212 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.106728 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:44Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.206096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.206141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.206155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.206174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.206186 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.309877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.310236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.310397 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.310603 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.310758 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.414053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.414096 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.414106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.414122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.414133 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.518368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.518447 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.518466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.518494 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.518540 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.629214 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.629284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.629305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.629331 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.629351 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.633616 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:44 crc kubenswrapper[4715]: E1203 21:42:44.633982 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.634821 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.634916 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.635028 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:44 crc kubenswrapper[4715]: E1203 21:42:44.635130 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:44 crc kubenswrapper[4715]: E1203 21:42:44.635026 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:44 crc kubenswrapper[4715]: E1203 21:42:44.635402 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.732311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.732392 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.732417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.732448 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.732475 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.835669 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.835744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.835768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.835798 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.835820 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.939082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.939122 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.939133 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.939150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:44 crc kubenswrapper[4715]: I1203 21:42:44.939161 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:44Z","lastTransitionTime":"2025-12-03T21:42:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.041801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.041853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.041870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.041894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.041917 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.145156 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.145237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.145255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.145280 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.145299 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.248631 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.248698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.248719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.248745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.248766 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.352650 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.352719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.352745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.352774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.352799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.456481 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.456567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.456587 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.456611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.456633 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.559400 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.559465 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.559485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.559547 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.559565 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.662851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.662911 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.662929 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.662953 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.662974 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.766391 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.766461 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.766479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.766548 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.766571 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.870103 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.870161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.870178 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.870202 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.870230 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.973567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.973635 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.973657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.973683 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:45 crc kubenswrapper[4715]: I1203 21:42:45.973703 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:45Z","lastTransitionTime":"2025-12-03T21:42:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.076729 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.076809 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.076835 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.076866 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.076892 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.180192 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.180268 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.180290 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.180322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.180347 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.309095 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.309153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.309356 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.309376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.309390 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.412764 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.412852 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.412871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.412899 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.412919 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.502704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.502768 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.502786 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.502815 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.502835 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.524535 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:46Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.530164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.530221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.530240 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.530263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.530280 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.550154 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:46Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.556048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.556111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.556128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.556155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.556172 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.586704 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:46Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.592801 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.592848 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.592865 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.592889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.592907 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.617157 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:46Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.622823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.622871 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.622889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.622913 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.622931 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.634126 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.634178 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.634283 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.634422 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.634473 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.635183 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.635339 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.635626 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.636033 4715 scope.go:117] "RemoveContainer" containerID="2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4" Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.636481 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.645658 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:46Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:46Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:46 crc kubenswrapper[4715]: E1203 21:42:46.645985 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.648590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.648643 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.648663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.648684 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.648700 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.656010 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.752181 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.752242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.752259 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.752287 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.752307 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.855490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.855579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.855601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.855629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.855650 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.961735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.962154 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.962306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.962444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:46 crc kubenswrapper[4715]: I1203 21:42:46.962599 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:46Z","lastTransitionTime":"2025-12-03T21:42:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.065562 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.065638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.065657 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.065682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.065700 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.169279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.169615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.169787 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.169957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.170189 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.273276 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.273327 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.273336 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.273353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.273366 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.376722 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.377069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.377207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.377381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.377539 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.480920 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.481130 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.481255 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.481384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.481537 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.586628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.587043 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.587185 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.587322 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.587449 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.690977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.691027 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.691045 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.691068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.691086 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.794419 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.794475 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.794493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.794553 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.794572 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.897780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.897838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.897857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.897880 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:47 crc kubenswrapper[4715]: I1203 21:42:47.897899 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:47Z","lastTransitionTime":"2025-12-03T21:42:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.000974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.001030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.001049 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.001081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.001100 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.103703 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.103761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.103783 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.103811 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.103832 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.208945 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.209074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.209091 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.209119 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.209139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.312636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.312701 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.312719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.312744 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.312766 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.415906 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.415959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.415977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.415999 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.416016 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.524141 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.524321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.524347 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.524424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.524494 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.628568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.628632 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.628687 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.628718 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.628738 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.634226 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.634254 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.634299 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.634352 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:48 crc kubenswrapper[4715]: E1203 21:42:48.634587 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:48 crc kubenswrapper[4715]: E1203 21:42:48.634849 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:48 crc kubenswrapper[4715]: E1203 21:42:48.635041 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:48 crc kubenswrapper[4715]: E1203 21:42:48.635198 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.732218 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.732288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.732312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.732343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.732371 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.837083 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.837155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.837172 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.837201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.837223 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.940705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.940761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.940780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.940806 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:48 crc kubenswrapper[4715]: I1203 21:42:48.940825 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:48Z","lastTransitionTime":"2025-12-03T21:42:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.044151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.044224 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.044247 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.044273 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.044293 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.147069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.147139 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.147159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.147186 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.147205 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.250320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.250384 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.250404 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.250430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.250449 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.354230 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.354297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.354316 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.354340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.354358 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.458022 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.458128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.458152 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.458189 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.458210 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.562353 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.562420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.562440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.562467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.562485 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.665595 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.665659 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.665679 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.665702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.665721 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.769342 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.769398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.769416 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.769440 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.769456 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.872606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.872709 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.872728 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.872753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.872773 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.975936 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.975995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.976012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.976039 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:49 crc kubenswrapper[4715]: I1203 21:42:49.976059 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:49Z","lastTransitionTime":"2025-12-03T21:42:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.079222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.079302 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.079325 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.079354 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.079375 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.182987 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.183077 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.183097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.183127 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.183146 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.285922 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.285984 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.286001 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.286026 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.286043 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.388886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.388959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.388983 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.389015 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.389037 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.492200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.492267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.492285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.492311 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.492330 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.595285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.595343 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.595361 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.595385 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.595402 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.634031 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.634082 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.634037 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:50 crc kubenswrapper[4715]: E1203 21:42:50.634208 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.634263 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:50 crc kubenswrapper[4715]: E1203 21:42:50.634447 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:50 crc kubenswrapper[4715]: E1203 21:42:50.634605 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:50 crc kubenswrapper[4715]: E1203 21:42:50.634862 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.698460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.698568 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.698592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.698624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.698647 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.801073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.801134 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.801151 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.801175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.801193 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.903542 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.903606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.903624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.903649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:50 crc kubenswrapper[4715]: I1203 21:42:50.903669 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:50Z","lastTransitionTime":"2025-12-03T21:42:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.006615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.006705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.006723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.006747 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.006766 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.109621 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.109672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.109689 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.109712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.109731 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.212374 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.212427 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.212441 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.212460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.212473 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.315117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.315194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.315211 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.315236 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.315253 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.418431 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.418493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.418536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.418561 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.418581 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.521851 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.521926 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.521949 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.521977 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.521994 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.625774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.625859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.625878 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.625901 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.625918 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.729048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.729105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.729123 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.729147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.729160 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.832376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.832434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.832455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.832479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.832496 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.935630 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.935884 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.935990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.936025 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:51 crc kubenswrapper[4715]: I1203 21:42:51.936093 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:51Z","lastTransitionTime":"2025-12-03T21:42:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.039985 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.040018 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.040029 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.040042 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.040052 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.143398 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.143467 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.143492 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.143558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.143579 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.246474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.246560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.246577 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.246599 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.246613 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.350367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.350409 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.350426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.350446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.350458 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.453153 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.453235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.453249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.453270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.453283 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.556069 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.556128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.556146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.556170 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.556188 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.634138 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.634146 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:52 crc kubenswrapper[4715]: E1203 21:42:52.634288 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.634171 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.634332 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:52 crc kubenswrapper[4715]: E1203 21:42:52.634438 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:52 crc kubenswrapper[4715]: E1203 21:42:52.634558 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:52 crc kubenswrapper[4715]: E1203 21:42:52.634683 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.659990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.660043 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.660055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.660075 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.660089 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.762978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.763051 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.763068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.763093 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.763114 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.866803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.866847 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.866864 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.866887 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.866904 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.969239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.969283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.969299 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.969321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:52 crc kubenswrapper[4715]: I1203 21:42:52.969337 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:52Z","lastTransitionTime":"2025-12-03T21:42:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.072661 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.072730 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.072756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.072784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.072804 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.175892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.175995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.176014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.176038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.176061 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.279050 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.279147 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.279168 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.279195 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.279212 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.382345 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.382418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.382443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.382472 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.382498 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.486297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.486362 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.486380 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.486403 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.486422 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.589425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.589469 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.589488 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.589540 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.589559 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.653363 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.672201 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.693053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.693193 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.693223 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.693254 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.693290 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.706471 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.723424 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.754316 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.771225 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.789050 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.799545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.799611 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.799681 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.799714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.799733 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.804748 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.821721 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.844201 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.860481 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49845bd6-cc63-4f3b-88ff-e1786256fa2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b212fc5927f3c43b22e22d9f3e58bc44022ee097b55548ce38d99f12341f1e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.880014 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.902292 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.903162 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.903191 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.903229 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.903249 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.903264 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:53Z","lastTransitionTime":"2025-12-03T21:42:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.925256 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.947706 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.966163 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:53 crc kubenswrapper[4715]: I1203 21:42:53.986860 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:53Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.007460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.007578 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.007636 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.007672 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.007731 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.008801 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:54Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.028688 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:54Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.112493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.112564 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.112582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.112609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.112627 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.216834 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.216915 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.216933 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.216960 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.216978 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.320592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.320663 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.320682 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.320706 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.320733 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.424097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.424201 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.424230 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.424267 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.424294 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.527532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.527598 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.527623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.527655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.527680 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.631433 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.631497 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.631541 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.631567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.631584 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.633717 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.633758 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.633798 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.633824 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:54 crc kubenswrapper[4715]: E1203 21:42:54.633957 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:54 crc kubenswrapper[4715]: E1203 21:42:54.634119 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:54 crc kubenswrapper[4715]: E1203 21:42:54.634236 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:54 crc kubenswrapper[4715]: E1203 21:42:54.634347 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.735111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.735175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.735194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.735221 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.735242 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.837914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.837972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.837995 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.838024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.838043 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.940723 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.940826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.940845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.940905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:54 crc kubenswrapper[4715]: I1203 21:42:54.940924 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:54Z","lastTransitionTime":"2025-12-03T21:42:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.044337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.044430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.044457 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.044496 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.044559 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.147306 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.147342 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.147351 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.147364 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.147374 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.249607 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.249688 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.249711 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.249745 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.249769 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.341697 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:55 crc kubenswrapper[4715]: E1203 21:42:55.341944 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:55 crc kubenswrapper[4715]: E1203 21:42:55.342105 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs podName:34ccf04b-c534-43bb-94ee-fc2e7cef68c6 nodeName:}" failed. No retries permitted until 2025-12-03 21:43:27.342071048 +0000 UTC m=+104.084781703 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs") pod "network-metrics-daemon-hbsmx" (UID: "34ccf04b-c534-43bb-94ee-fc2e7cef68c6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.352367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.352437 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.352462 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.352534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.352562 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.455743 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.456274 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.456292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.456318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.456336 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.559477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.559567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.559585 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.559615 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.559633 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.661793 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.661853 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.661870 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.661892 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.661910 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.765150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.765222 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.765245 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.765277 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.765309 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.869854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.869921 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.869944 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.869974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.869996 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.972182 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.972243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.972261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.972283 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:55 crc kubenswrapper[4715]: I1203 21:42:55.972299 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:55Z","lastTransitionTime":"2025-12-03T21:42:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.075914 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.075971 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.075989 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.076011 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.076028 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.120287 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/0.log" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.120368 4715 generic.go:334] "Generic (PLEG): container finished" podID="9174d88f-f7f0-4b0c-942f-39d027b97e34" containerID="0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848" exitCode=1 Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.120411 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2wcn9" event={"ID":"9174d88f-f7f0-4b0c-942f-39d027b97e34","Type":"ContainerDied","Data":"0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.120983 4715 scope.go:117] "RemoveContainer" containerID="0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.144179 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.170326 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.179974 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.180038 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.180056 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.180109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.180126 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.187652 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.199426 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49845bd6-cc63-4f3b-88ff-e1786256fa2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b212fc5927f3c43b22e22d9f3e58bc44022ee097b55548ce38d99f12341f1e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.220603 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.240222 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.256070 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.278904 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.282671 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.282727 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.282746 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.282773 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.282792 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.297117 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.315903 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.333367 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.363274 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.384253 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.386756 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.386820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.386841 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.387068 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.387110 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.407393 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:55Z\\\",\\\"message\\\":\\\"2025-12-03T21:42:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c\\\\n2025-12-03T21:42:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c to /host/opt/cni/bin/\\\\n2025-12-03T21:42:10Z [verbose] multus-daemon started\\\\n2025-12-03T21:42:10Z [verbose] Readiness Indicator file check\\\\n2025-12-03T21:42:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.442169 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.461987 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.491381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.491455 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.491479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.491555 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.491576 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.494405 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.514169 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.533384 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.595426 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.595487 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.595532 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.595558 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.595576 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.634148 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.634196 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.634297 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.634497 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.634576 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.634698 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.634848 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.635100 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.698976 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.699044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.699062 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.699092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.699114 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.802261 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.802334 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.802352 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.802376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.802394 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.843845 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.843909 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.843927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.843952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.843972 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.869336 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.875459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.875531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.875550 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.875569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.875583 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.891824 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.895739 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.895766 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.895779 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.895797 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.895809 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.913453 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.918206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.918235 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.918243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.918260 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.918269 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.935791 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.941053 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.941094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.941107 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.941126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.941139 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.961759 4715 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"cb36c1b6-a2af-4152-9c42-bbfe6dbddc4f\\\",\\\"systemUUID\\\":\\\"14ababa8-2b39-4191-882f-3b54d73698cb\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:56Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:56 crc kubenswrapper[4715]: E1203 21:42:56.962084 4715 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.964506 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.964606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.964626 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.964656 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:56 crc kubenswrapper[4715]: I1203 21:42:56.964673 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:56Z","lastTransitionTime":"2025-12-03T21:42:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.068534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.068602 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.068620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.068649 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.068668 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.127874 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/0.log" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.127972 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2wcn9" event={"ID":"9174d88f-f7f0-4b0c-942f-39d027b97e34","Type":"ContainerStarted","Data":"d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.148116 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49845bd6-cc63-4f3b-88ff-e1786256fa2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b212fc5927f3c43b22e22d9f3e58bc44022ee097b55548ce38d99f12341f1e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.171712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.171695 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.171792 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.171820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.171854 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.171879 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.195371 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.214890 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.243404 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.267956 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.274774 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.274879 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.274908 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.274948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.274980 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.291218 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.317745 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.339118 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.365758 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:55Z\\\",\\\"message\\\":\\\"2025-12-03T21:42:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c\\\\n2025-12-03T21:42:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c to /host/opt/cni/bin/\\\\n2025-12-03T21:42:10Z [verbose] multus-daemon started\\\\n2025-12-03T21:42:10Z [verbose] Readiness Indicator file check\\\\n2025-12-03T21:42:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.378990 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.379052 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.379070 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.379112 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.379134 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.386041 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.457870 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.482340 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.482381 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.482393 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.482412 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.482426 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.485325 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.512210 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.527288 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.541367 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.558358 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.576400 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.585875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.585938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.585956 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.585978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.585994 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.591778 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:42:57Z is after 2025-08-24T17:21:41Z" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.688856 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.688924 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.688941 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.688972 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.688990 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.791275 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.791339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.791358 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.791386 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.791405 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.895084 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.895144 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.895161 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.895187 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.895204 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.998466 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.998567 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.998592 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.998623 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:57 crc kubenswrapper[4715]: I1203 21:42:57.998646 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:57Z","lastTransitionTime":"2025-12-03T21:42:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.101875 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.101935 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.101957 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.101980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.101999 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.207145 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.207230 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.207251 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.207284 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.207309 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.311257 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.311320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.311337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.311367 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.311386 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.415337 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.415803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.415962 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.416109 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.416261 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.520155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.520237 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.520256 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.520279 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.520294 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.623446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.623677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.623705 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.623735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.623753 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.634208 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.634206 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.634217 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.634290 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:42:58 crc kubenswrapper[4715]: E1203 21:42:58.634854 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:42:58 crc kubenswrapper[4715]: E1203 21:42:58.634965 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:42:58 crc kubenswrapper[4715]: E1203 21:42:58.635062 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:42:58 crc kubenswrapper[4715]: E1203 21:42:58.635143 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.727323 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.727388 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.727407 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.727432 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.727452 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.830889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.830952 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.830970 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.830991 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.831008 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.934014 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.934092 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.934110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.934146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:58 crc kubenswrapper[4715]: I1203 21:42:58.934167 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:58Z","lastTransitionTime":"2025-12-03T21:42:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.037741 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.037789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.037802 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.037823 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.037837 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.140943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.141031 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.141054 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.141085 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.141105 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.244443 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.244569 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.244625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.244654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.244836 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.348207 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.348278 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.348297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.348324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.348343 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.451537 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.451600 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.451625 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.451647 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.451664 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.555212 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.555297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.555317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.555349 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.555369 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.658200 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.658270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.658288 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.658318 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.658336 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.761373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.761439 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.761459 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.761485 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.761542 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.864155 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.864225 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.864242 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.864272 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.864292 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.967629 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.967702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.967721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.967749 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:42:59 crc kubenswrapper[4715]: I1203 21:42:59.967770 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:42:59Z","lastTransitionTime":"2025-12-03T21:42:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.070030 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.070095 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.070114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.070140 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.070158 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.173305 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.173373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.173394 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.173424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.173444 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.277117 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.277434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.277470 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.277531 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.277804 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.380973 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.381057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.381080 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.381110 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.381131 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.484324 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.484434 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.484468 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.484534 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.484561 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.588243 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.588321 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.588339 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.588368 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.588390 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.634216 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.634281 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.634399 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.634646 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:00 crc kubenswrapper[4715]: E1203 21:43:00.634643 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:00 crc kubenswrapper[4715]: E1203 21:43:00.634791 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:00 crc kubenswrapper[4715]: E1203 21:43:00.634928 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:00 crc kubenswrapper[4715]: E1203 21:43:00.635143 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.636241 4715 scope.go:117] "RemoveContainer" containerID="2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.691997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.692061 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.692082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.692111 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.692131 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.795824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.795918 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.795938 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.796003 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.796023 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.900206 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.900270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.900292 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.900320 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:00 crc kubenswrapper[4715]: I1203 21:43:00.900339 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:00Z","lastTransitionTime":"2025-12-03T21:43:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.003639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.003704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.003721 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.003751 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.003769 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.107344 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.107421 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.107444 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.107479 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.107543 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.148536 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/2.log" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.152271 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.152848 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.210063 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.210226 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.210271 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.210285 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.210312 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.210333 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.234782 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.250731 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.278711 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.304829 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49845bd6-cc63-4f3b-88ff-e1786256fa2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b212fc5927f3c43b22e22d9f3e58bc44022ee097b55548ce38d99f12341f1e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.313476 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.313572 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.313594 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.313622 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.313642 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.329143 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.344447 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.364700 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.382913 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:55Z\\\",\\\"message\\\":\\\"2025-12-03T21:42:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c\\\\n2025-12-03T21:42:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c to /host/opt/cni/bin/\\\\n2025-12-03T21:42:10Z [verbose] multus-daemon started\\\\n2025-12-03T21:42:10Z [verbose] Readiness Indicator file check\\\\n2025-12-03T21:42:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.399946 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.417817 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.417859 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.417872 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.417894 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.417908 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.418248 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.444553 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.480085 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.505260 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.525761 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.525812 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.525827 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.526048 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.526070 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.530328 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.567015 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.587206 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.603592 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.629940 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.630008 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.630029 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.630060 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.630085 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.631684 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:01Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.733539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.733948 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.734097 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.734248 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.734374 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.854373 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.854460 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.854480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.854536 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.854556 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.959029 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.959106 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.959126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.959159 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:01 crc kubenswrapper[4715]: I1203 21:43:01.959181 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:01Z","lastTransitionTime":"2025-12-03T21:43:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.062925 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.063009 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.063028 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.063058 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.063079 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.166780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.166857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.166877 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.166905 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.166928 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.270490 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.270601 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.270624 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.270654 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.270674 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.376128 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.376196 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.376215 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.376252 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.376273 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.480308 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.480800 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.481000 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.481150 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.481309 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.585493 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.585583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.585606 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.585638 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.585660 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.633272 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.633322 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.633299 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:02 crc kubenswrapper[4715]: E1203 21:43:02.633597 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:02 crc kubenswrapper[4715]: E1203 21:43:02.633442 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:02 crc kubenswrapper[4715]: E1203 21:43:02.633835 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.634615 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:02 crc kubenswrapper[4715]: E1203 21:43:02.634990 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.689355 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.689410 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.689424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.689446 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.689463 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.792919 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.792978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.792997 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.793021 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.793040 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.895799 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.895869 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.895888 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.895917 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.895937 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.998912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.998979 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.998998 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.999024 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:02 crc kubenswrapper[4715]: I1203 21:43:02.999048 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:02Z","lastTransitionTime":"2025-12-03T21:43:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.103480 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.103590 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.103610 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.103639 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.103661 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.163913 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/3.log" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.165138 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/2.log" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.169590 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" exitCode=1 Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.169655 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.169717 4715 scope.go:117] "RemoveContainer" containerID="2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.171540 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 21:43:03 crc kubenswrapper[4715]: E1203 21:43:03.171987 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.204708 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.207424 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.207482 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.207533 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.207560 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.207579 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.229067 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.248223 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.268184 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49845bd6-cc63-4f3b-88ff-e1786256fa2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b212fc5927f3c43b22e22d9f3e58bc44022ee097b55548ce38d99f12341f1e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.291570 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.310702 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.310784 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.310803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.310833 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.310856 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.314333 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.331184 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.357911 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.382659 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.404696 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.414102 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.414174 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.414194 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.414219 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.414238 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.425830 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.447862 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.471668 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:55Z\\\",\\\"message\\\":\\\"2025-12-03T21:42:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c\\\\n2025-12-03T21:42:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c to /host/opt/cni/bin/\\\\n2025-12-03T21:42:10Z [verbose] multus-daemon started\\\\n2025-12-03T21:42:10Z [verbose] Readiness Indicator file check\\\\n2025-12-03T21:42:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.491039 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.517604 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.517666 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.517686 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.517712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.517734 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.517985 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.539608 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.573701 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:43:02Z\\\",\\\"message\\\":\\\"etry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc after 0 failed attempt(s)\\\\nI1203 21:43:01.777915 6653 services_controller.go:356] Processing sync for service openshift-dns-operator/metrics for network=default\\\\nI1203 21:43:01.777917 6653 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1203 21:43:01.777134 6653 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 21:43:01.778123 6653 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.594189 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.613447 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.620731 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.620794 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.620813 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.620838 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.620860 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.659031 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://20c6cd1d6d19c20d4584843a854c573594f796712e6be67bcb71617eeea4a6ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.682603 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2wcn9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9174d88f-f7f0-4b0c-942f-39d027b97e34\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:55Z\\\",\\\"message\\\":\\\"2025-12-03T21:42:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c\\\\n2025-12-03T21:42:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a2ab3f3-6d2b-4115-8de1-1e6db100075c to /host/opt/cni/bin/\\\\n2025-12-03T21:42:10Z [verbose] multus-daemon started\\\\n2025-12-03T21:42:10Z [verbose] Readiness Indicator file check\\\\n2025-12-03T21:42:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b24hf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2wcn9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.706078 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ca0ce89-c3ec-421f-9954-c594760b16a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3576a14b23530a01d1c6bb568d3ec2eb491ae9943db34c119da8339024046d44\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k26gp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-ctls2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.724160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.724239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.724270 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.724300 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.724327 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.733290 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b081691-dd32-48ab-a8a5-720850285986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"et denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 21:42:02.004814 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 21:42:02.004843 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1203 21:42:02.004846 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1203 21:42:02.004851 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 21:42:02.004868 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 21:42:02.004877 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 21:42:02.004883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 21:42:02.004887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 21:42:02.013882 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-2058794696/tls.crt::/tmp/serving-cert-2058794696/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764798105\\\\\\\\\\\\\\\" (2025-12-03 21:41:45 +0000 UTC to 2026-01-02 21:41:46 +0000 UTC (now=2025-12-03 21:42:02.01383992 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014080 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764798116\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764798116\\\\\\\\\\\\\\\" (2025-12-03 20:41:56 +0000 UTC to 2026-12-03 20:41:56 +0000 UTC (now=2025-12-03 21:42:02.014052375 +0000 UTC))\\\\\\\"\\\\nI1203 21:42:02.014103 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1203 21:42:02.014131 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nF1203 21:42:02.014444 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.758093 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"edd76bf9-d9a5-470e-a8cd-93b9780e6ede\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f59a4a5721883cab2c15bfe3b2e8a7765e85a643d2d808e6da5284537f392372\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://997b0e2ee83c09e2f5be4fc05a81fedd02d6b195b018114ab607c3dc19e3eac7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://02ead1f5f1411ddf137ec593caef0b938dabd2a7b9a20d1cc823c40e19ee57fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.782039 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"52fef3b0-2b28-46bf-a7b5-6e3f910b4192\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dd8273b18d9f8b38f97e55191172d16c3dc67cc2018271572b936a49db9a85f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0f193862b69be38266cdcf20bc435133507966bef4ff1379dd689e1d9408297d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48a3c9179422784de9c6df012962b16384c4bf12242fbdb5b40ad5e078a8b9b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f1850b8c764b4bee63bf4ce5ba3d31a196bab498917e3e21f45044a86c8a0073\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.803592 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea6cedd1-c3cc-4408-b6f6-a58fcac998a9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6339b42304f4c622a27aa4075b738a96a108d875df37e80b45f15d2d65e4534b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca52eee639794de836ba00472cd560206a055c8e4103fa05fe6b721143272ecb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bljbm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-47tx8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.822561 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:23Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dg2g5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:23Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-hbsmx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.828820 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.828912 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.828943 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.828978 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.829003 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.858646 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"511c9bf0-9554-4243-a7a9-4ad5dd602a15\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e6ec4953a630b4d9e9a034a5f620b7a20543f9bcea5a3eb399e5a4080817e7ef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8259e519d5e7c1bbf4e991ff939ad98227c145492adeabd03b1f88ffe98c94a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5fcbf5d7b810bcd4b8766c0e68f7aac4d80484e499f2e8badef64bc127662ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fff93a8517bd63af2e61d3ebc445837ae54e9ed9c907e3a6cd25ebed7aacec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://44de7604f74cc9bf07d17c4f46457be4ce5cc192d298578397aa9081c9607c70\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aab8b0e06d72c320d2f3a4d234248c0f8de7c550eaf560818551bae91f68d7cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1dd0dbf3603d6a14cf4867f99321294ecb12404e08b7f77ff00f0d5e509b393a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47c133d71f9ebe567bb307c53ca720bbebf85820ef67f83b0815b236a299bc63\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.904786 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:05Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://67bef0cde1662fa55cffd29a3b869c5b0de0954ce132b57aeb81668ca143aaa5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.932657 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ba9d6059-9d98-4816-8fad-9e430e516d70\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2825fcbb30a3ada2e7ca62e90201aae2b47a495992d59a43e2ddc7e5965b9dc4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:42:34Z\\\",\\\"message\\\":\\\"ts:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Groups:[]string{}}, services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_switch_crc\\\\\\\", UUID:\\\\\\\"b21188fe-5483-4717-afe6-20a41a40b91a\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{\\\\\\\"crc\\\\\\\"}, Routers:[]string{}, Groups:[]string{}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_default/kubernetes_TCP_node_router_crc\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"default/kubernetes\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.1\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{services.Addr{IP:\\\\\\\"169.254.0.2\\\\\\\", Port:6443, Template:(*services.Template)(nil)}}}}, Templates:services.TemplateMap(nil), Switches:[]string(nil), Routers:[]string{\\\\\\\"GR_crc\\\\\\\"}, Grou\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:33Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T21:43:02Z\\\",\\\"message\\\":\\\"etry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc after 0 failed attempt(s)\\\\nI1203 21:43:01.777915 6653 services_controller.go:356] Processing sync for service openshift-dns-operator/metrics for network=default\\\\nI1203 21:43:01.777917 6653 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/kube-rbac-proxy-crio-crc\\\\nI1203 21:43:01.777134 6653 model_client.go:382] Update operations generated as: [{Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.59 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {dce28c51-c9f1-478b-97c8-7e209d6e7cbe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1203 21:43:01.778123 6653 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurr\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T21:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xpm4j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-8j8wk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.933076 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.933163 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.933176 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.933220 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.933244 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:03Z","lastTransitionTime":"2025-12-03T21:43:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.948855 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.967662 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:03 crc kubenswrapper[4715]: I1203 21:43:03.994617 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7sg46" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fc36a21-545b-4055-92ae-6149beb86b89\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1ad68b0edcdbe4936a0951b1046b3fbff8436404812f7daaa15e1c0f2b58e4cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgcb9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:13Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7sg46\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:03Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.013953 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9wz2q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af42f462-12a2-4488-b408-8753ac62bcf4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72c70d33af22e603f04cd154fcbce244425e8242b53c03dcce2f59ea10298705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gtwvp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9wz2q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:04Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.035084 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5625d33c-6de7-4413-b9a5-e3ec2df2ef98\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://379c0f4c3afb9a85d543a4e76c64790a470f79b6addf4777a6ab7dda18b7dde1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://460eee3d31133321d466146be2bbcd7d83b8093e564459685476a3c8441c27e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9955d8369de9424159746f384fa7edf39785dbc8a460b504aea3f37f1d91883b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7feec5fa5f63478c25fa3c6d7e7ab2f5765699881a32422cd0c87b91819b49a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://db3cad276df55d5316443f649f0653ae4972b406d6629cf067d153e1cd00cbd7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5bebfcd48d17ef3d03b66361856decae51351fe888c1ff0f591ea43ca10f150\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b5d4b6c0212af56da8208a77e3cfe242d5869181eca5ddf79fe6b3c470d6c371\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:42:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wzbc2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:42:09Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-wmhrj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:04Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.036034 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.036082 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.036094 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.036113 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.036126 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.050933 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"49845bd6-cc63-4f3b-88ff-e1786256fa2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T21:41:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b212fc5927f3c43b22e22d9f3e58bc44022ee097b55548ce38d99f12341f1e74\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5bb8f534109a6d44fb92cc2d4399d33e23793c1aa96a40e495ab17273d735fae\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T21:41:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T21:41:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T21:41:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:04Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.072233 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:03Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3286ac3adcc6afcf3648e4662da72b383f9656b34e66ea582a73e71296926472\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cac72f1610f437c6f82bf90443fc3d8e186453abf4e4d38c0d080db898954add\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T21:42:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:04Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.091676 4715 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T21:42:02Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T21:43:04Z is after 2025-08-24T17:21:41Z" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.139982 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.140055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.140074 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.140105 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.140123 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.178177 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/3.log" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.242770 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.242840 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.242857 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.242881 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.242920 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.346477 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.346589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.346609 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.346698 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.346718 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.450073 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.450146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.450164 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.450189 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.450208 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.553695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.553742 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.553758 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.553780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.553799 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.634050 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.634171 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:04 crc kubenswrapper[4715]: E1203 21:43:04.634364 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.634412 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.634544 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:04 crc kubenswrapper[4715]: E1203 21:43:04.634695 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:04 crc kubenswrapper[4715]: E1203 21:43:04.635052 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:04 crc kubenswrapper[4715]: E1203 21:43:04.635819 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.657160 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.657239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.657265 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.657296 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.657320 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.760735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.760803 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.760826 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.760855 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.760875 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.864788 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.864843 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.864862 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.864886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.864907 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.968425 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.968539 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.968559 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.968589 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:04 crc kubenswrapper[4715]: I1203 21:43:04.968609 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:04Z","lastTransitionTime":"2025-12-03T21:43:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.072474 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.072574 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.072593 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.072620 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.072638 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.176420 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.176498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.176551 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.176582 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.176603 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.280263 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.280612 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.281089 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.281269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.281401 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.384704 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.384778 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.384795 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.384824 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.384844 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.488197 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.488269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.488289 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.488317 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.488340 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.592583 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.592655 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.592673 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.592700 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.592720 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.696086 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.696217 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.696239 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.696269 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.696289 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.801580 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.801634 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.801648 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.801665 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.801680 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.905735 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.906297 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.906545 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.906712 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:05 crc kubenswrapper[4715]: I1203 21:43:05.906854 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:05Z","lastTransitionTime":"2025-12-03T21:43:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.010789 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.010844 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.010863 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.010889 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.010908 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.113928 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.114012 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.114032 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.114055 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.114071 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.216886 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.216959 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.216980 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.217006 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.217029 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.321058 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.321126 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.321146 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.321175 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.321201 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.443350 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.443413 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.443430 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.443456 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.443473 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.486775 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.486927 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.486897147 +0000 UTC m=+147.229607782 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.487033 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.487087 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.487133 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.487193 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487268 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487300 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487319 4715 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487354 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487358 4715 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487384 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.487363109 +0000 UTC m=+147.230073744 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487431 4715 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487472 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.487438661 +0000 UTC m=+147.230149286 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487378 4715 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487535 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.487491183 +0000 UTC m=+147.230201808 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487544 4715 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.487596 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.487580776 +0000 UTC m=+147.230291401 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.546986 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.547057 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.547081 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.547114 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.547142 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.633810 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.633895 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.633924 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.633832 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.634048 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.634167 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.634268 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:06 crc kubenswrapper[4715]: E1203 21:43:06.634351 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.651326 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.651376 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.651395 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.651417 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.651437 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.755628 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.755677 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.755695 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.755719 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.755737 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.859418 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.859498 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.859550 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.859579 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.859599 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.963927 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.963993 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.964013 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.964044 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:06 crc kubenswrapper[4715]: I1203 21:43:06.964064 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:06Z","lastTransitionTime":"2025-12-03T21:43:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.014658 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.014734 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.014753 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.014780 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.014800 4715 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T21:43:07Z","lastTransitionTime":"2025-12-03T21:43:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.094734 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w"] Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.095587 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.100059 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.100015 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.100158 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.100675 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.144091 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=65.144059689 podStartE2EDuration="1m5.144059689s" podCreationTimestamp="2025-12-03 21:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.142665432 +0000 UTC m=+83.885376057" watchObservedRunningTime="2025-12-03 21:43:07.144059689 +0000 UTC m=+83.886770284" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.196493 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dc714618-33d4-4149-98e9-9f915ab8928c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.196636 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc714618-33d4-4149-98e9-9f915ab8928c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.196676 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc714618-33d4-4149-98e9-9f915ab8928c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.196737 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dc714618-33d4-4149-98e9-9f915ab8928c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.196816 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc714618-33d4-4149-98e9-9f915ab8928c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.234246 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-47tx8" podStartSLOduration=58.234208439 podStartE2EDuration="58.234208439s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.233983923 +0000 UTC m=+83.976694558" watchObservedRunningTime="2025-12-03 21:43:07.234208439 +0000 UTC m=+83.976919074" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.298679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dc714618-33d4-4149-98e9-9f915ab8928c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.298774 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc714618-33d4-4149-98e9-9f915ab8928c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.298809 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc714618-33d4-4149-98e9-9f915ab8928c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.298875 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dc714618-33d4-4149-98e9-9f915ab8928c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.298952 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc714618-33d4-4149-98e9-9f915ab8928c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.299492 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/dc714618-33d4-4149-98e9-9f915ab8928c-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.301933 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/dc714618-33d4-4149-98e9-9f915ab8928c-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.306959 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/dc714618-33d4-4149-98e9-9f915ab8928c-service-ca\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.322633 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc714618-33d4-4149-98e9-9f915ab8928c-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.334083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc714618-33d4-4149-98e9-9f915ab8928c-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-pnm6w\" (UID: \"dc714618-33d4-4149-98e9-9f915ab8928c\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.350428 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7sg46" podStartSLOduration=59.350398202 podStartE2EDuration="59.350398202s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.349861987 +0000 UTC m=+84.092572622" watchObservedRunningTime="2025-12-03 21:43:07.350398202 +0000 UTC m=+84.093108807" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.383330 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.383305587 podStartE2EDuration="21.383305587s" podCreationTimestamp="2025-12-03 21:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.363465709 +0000 UTC m=+84.106176314" watchObservedRunningTime="2025-12-03 21:43:07.383305587 +0000 UTC m=+84.126016192" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.417351 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9wz2q" podStartSLOduration=59.417320043 podStartE2EDuration="59.417320043s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.416481471 +0000 UTC m=+84.159192096" watchObservedRunningTime="2025-12-03 21:43:07.417320043 +0000 UTC m=+84.160030648" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.418782 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.452949 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-wmhrj" podStartSLOduration=59.452924371 podStartE2EDuration="59.452924371s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.452601822 +0000 UTC m=+84.195312427" watchObservedRunningTime="2025-12-03 21:43:07.452924371 +0000 UTC m=+84.195634976" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.480183 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=65.480163975 podStartE2EDuration="1m5.480163975s" podCreationTimestamp="2025-12-03 21:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.478740447 +0000 UTC m=+84.221451052" watchObservedRunningTime="2025-12-03 21:43:07.480163975 +0000 UTC m=+84.222874580" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.521443 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=65.521420614 podStartE2EDuration="1m5.521420614s" podCreationTimestamp="2025-12-03 21:42:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.498676199 +0000 UTC m=+84.241386804" watchObservedRunningTime="2025-12-03 21:43:07.521420614 +0000 UTC m=+84.264131219" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.542280 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=37.542266129 podStartE2EDuration="37.542266129s" podCreationTimestamp="2025-12-03 21:42:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.523000986 +0000 UTC m=+84.265711601" watchObservedRunningTime="2025-12-03 21:43:07.542266129 +0000 UTC m=+84.284976764" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.562976 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2wcn9" podStartSLOduration=59.56296435 podStartE2EDuration="59.56296435s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.56072539 +0000 UTC m=+84.303436015" watchObservedRunningTime="2025-12-03 21:43:07.56296435 +0000 UTC m=+84.305674975" Dec 03 21:43:07 crc kubenswrapper[4715]: I1203 21:43:07.576418 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podStartSLOduration=59.576366926 podStartE2EDuration="59.576366926s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:07.574956069 +0000 UTC m=+84.317666664" watchObservedRunningTime="2025-12-03 21:43:07.576366926 +0000 UTC m=+84.319077531" Dec 03 21:43:08 crc kubenswrapper[4715]: I1203 21:43:08.204924 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" event={"ID":"dc714618-33d4-4149-98e9-9f915ab8928c","Type":"ContainerStarted","Data":"9ba4056b00186407135228ed31dca778780b24e5b5fd02a24cacd71f28d1bc79"} Dec 03 21:43:08 crc kubenswrapper[4715]: I1203 21:43:08.633454 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:08 crc kubenswrapper[4715]: I1203 21:43:08.633542 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:08 crc kubenswrapper[4715]: I1203 21:43:08.633638 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:08 crc kubenswrapper[4715]: E1203 21:43:08.633786 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:08 crc kubenswrapper[4715]: I1203 21:43:08.633850 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:08 crc kubenswrapper[4715]: E1203 21:43:08.633996 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:08 crc kubenswrapper[4715]: E1203 21:43:08.634232 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:08 crc kubenswrapper[4715]: E1203 21:43:08.634321 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:09 crc kubenswrapper[4715]: I1203 21:43:09.211294 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" event={"ID":"dc714618-33d4-4149-98e9-9f915ab8928c","Type":"ContainerStarted","Data":"7d7619cc4a0e957b202e98c4723df5c2333f55de58489b11cbb5864ac6a9e479"} Dec 03 21:43:10 crc kubenswrapper[4715]: I1203 21:43:10.633885 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:10 crc kubenswrapper[4715]: I1203 21:43:10.633999 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:10 crc kubenswrapper[4715]: I1203 21:43:10.633937 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:10 crc kubenswrapper[4715]: I1203 21:43:10.633925 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:10 crc kubenswrapper[4715]: E1203 21:43:10.634148 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:10 crc kubenswrapper[4715]: E1203 21:43:10.634350 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:10 crc kubenswrapper[4715]: E1203 21:43:10.634485 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:10 crc kubenswrapper[4715]: E1203 21:43:10.634613 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:12 crc kubenswrapper[4715]: I1203 21:43:12.633665 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:12 crc kubenswrapper[4715]: I1203 21:43:12.633719 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:12 crc kubenswrapper[4715]: I1203 21:43:12.633731 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:12 crc kubenswrapper[4715]: E1203 21:43:12.634214 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:12 crc kubenswrapper[4715]: E1203 21:43:12.634035 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:12 crc kubenswrapper[4715]: I1203 21:43:12.633829 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:12 crc kubenswrapper[4715]: E1203 21:43:12.634451 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:12 crc kubenswrapper[4715]: E1203 21:43:12.634579 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:14 crc kubenswrapper[4715]: I1203 21:43:14.633864 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:14 crc kubenswrapper[4715]: I1203 21:43:14.633923 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:14 crc kubenswrapper[4715]: I1203 21:43:14.634123 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:14 crc kubenswrapper[4715]: E1203 21:43:14.634360 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:14 crc kubenswrapper[4715]: E1203 21:43:14.634540 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:14 crc kubenswrapper[4715]: E1203 21:43:14.635092 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:14 crc kubenswrapper[4715]: I1203 21:43:14.635400 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:14 crc kubenswrapper[4715]: I1203 21:43:14.635574 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 21:43:14 crc kubenswrapper[4715]: E1203 21:43:14.635846 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" Dec 03 21:43:14 crc kubenswrapper[4715]: E1203 21:43:14.635895 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:14 crc kubenswrapper[4715]: I1203 21:43:14.688064 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-pnm6w" podStartSLOduration=66.688034631 podStartE2EDuration="1m6.688034631s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:09.233799973 +0000 UTC m=+85.976510608" watchObservedRunningTime="2025-12-03 21:43:14.688034631 +0000 UTC m=+91.430745256" Dec 03 21:43:16 crc kubenswrapper[4715]: I1203 21:43:16.633675 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:16 crc kubenswrapper[4715]: I1203 21:43:16.633719 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:16 crc kubenswrapper[4715]: I1203 21:43:16.633776 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:16 crc kubenswrapper[4715]: I1203 21:43:16.633687 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:16 crc kubenswrapper[4715]: E1203 21:43:16.633871 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:16 crc kubenswrapper[4715]: E1203 21:43:16.633999 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:16 crc kubenswrapper[4715]: E1203 21:43:16.634225 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:16 crc kubenswrapper[4715]: E1203 21:43:16.634417 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:18 crc kubenswrapper[4715]: I1203 21:43:18.634236 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:18 crc kubenswrapper[4715]: I1203 21:43:18.634427 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:18 crc kubenswrapper[4715]: E1203 21:43:18.634550 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:18 crc kubenswrapper[4715]: I1203 21:43:18.634589 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:18 crc kubenswrapper[4715]: E1203 21:43:18.635098 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:18 crc kubenswrapper[4715]: E1203 21:43:18.635260 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:18 crc kubenswrapper[4715]: I1203 21:43:18.636015 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:18 crc kubenswrapper[4715]: E1203 21:43:18.636359 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:20 crc kubenswrapper[4715]: I1203 21:43:20.633963 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:20 crc kubenswrapper[4715]: I1203 21:43:20.633977 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:20 crc kubenswrapper[4715]: I1203 21:43:20.634117 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:20 crc kubenswrapper[4715]: I1203 21:43:20.634715 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:20 crc kubenswrapper[4715]: E1203 21:43:20.636113 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:20 crc kubenswrapper[4715]: E1203 21:43:20.636336 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:20 crc kubenswrapper[4715]: E1203 21:43:20.636694 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:20 crc kubenswrapper[4715]: E1203 21:43:20.636910 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:22 crc kubenswrapper[4715]: I1203 21:43:22.634080 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:22 crc kubenswrapper[4715]: E1203 21:43:22.634188 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:22 crc kubenswrapper[4715]: I1203 21:43:22.634384 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:22 crc kubenswrapper[4715]: E1203 21:43:22.634429 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:22 crc kubenswrapper[4715]: I1203 21:43:22.634739 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:22 crc kubenswrapper[4715]: E1203 21:43:22.634788 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:22 crc kubenswrapper[4715]: I1203 21:43:22.634971 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:22 crc kubenswrapper[4715]: E1203 21:43:22.635209 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:24 crc kubenswrapper[4715]: I1203 21:43:24.633340 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:24 crc kubenswrapper[4715]: I1203 21:43:24.633580 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:24 crc kubenswrapper[4715]: I1203 21:43:24.633691 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:24 crc kubenswrapper[4715]: I1203 21:43:24.633737 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:24 crc kubenswrapper[4715]: E1203 21:43:24.633613 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:24 crc kubenswrapper[4715]: E1203 21:43:24.633829 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:24 crc kubenswrapper[4715]: E1203 21:43:24.633880 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:24 crc kubenswrapper[4715]: E1203 21:43:24.633992 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:26 crc kubenswrapper[4715]: I1203 21:43:26.633976 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:26 crc kubenswrapper[4715]: I1203 21:43:26.634063 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:26 crc kubenswrapper[4715]: I1203 21:43:26.634120 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:26 crc kubenswrapper[4715]: I1203 21:43:26.634012 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:26 crc kubenswrapper[4715]: E1203 21:43:26.634250 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:26 crc kubenswrapper[4715]: E1203 21:43:26.634414 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:26 crc kubenswrapper[4715]: E1203 21:43:26.634606 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:26 crc kubenswrapper[4715]: E1203 21:43:26.634693 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:27 crc kubenswrapper[4715]: I1203 21:43:27.356116 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:27 crc kubenswrapper[4715]: E1203 21:43:27.356452 4715 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:43:27 crc kubenswrapper[4715]: E1203 21:43:27.356658 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs podName:34ccf04b-c534-43bb-94ee-fc2e7cef68c6 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:31.356618556 +0000 UTC m=+168.099329181 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs") pod "network-metrics-daemon-hbsmx" (UID: "34ccf04b-c534-43bb-94ee-fc2e7cef68c6") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 21:43:28 crc kubenswrapper[4715]: I1203 21:43:28.633360 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:28 crc kubenswrapper[4715]: I1203 21:43:28.633411 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:28 crc kubenswrapper[4715]: I1203 21:43:28.633487 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:28 crc kubenswrapper[4715]: I1203 21:43:28.633592 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:28 crc kubenswrapper[4715]: E1203 21:43:28.633724 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:28 crc kubenswrapper[4715]: E1203 21:43:28.633957 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:28 crc kubenswrapper[4715]: E1203 21:43:28.634366 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:28 crc kubenswrapper[4715]: E1203 21:43:28.634490 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:29 crc kubenswrapper[4715]: I1203 21:43:29.636483 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 21:43:29 crc kubenswrapper[4715]: E1203 21:43:29.636961 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-8j8wk_openshift-ovn-kubernetes(ba9d6059-9d98-4816-8fad-9e430e516d70)\"" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" Dec 03 21:43:30 crc kubenswrapper[4715]: I1203 21:43:30.634369 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:30 crc kubenswrapper[4715]: I1203 21:43:30.634450 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:30 crc kubenswrapper[4715]: I1203 21:43:30.634370 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:30 crc kubenswrapper[4715]: E1203 21:43:30.634614 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:30 crc kubenswrapper[4715]: I1203 21:43:30.634705 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:30 crc kubenswrapper[4715]: E1203 21:43:30.634976 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:30 crc kubenswrapper[4715]: E1203 21:43:30.635087 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:30 crc kubenswrapper[4715]: E1203 21:43:30.635210 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:32 crc kubenswrapper[4715]: I1203 21:43:32.634099 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:32 crc kubenswrapper[4715]: I1203 21:43:32.634138 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:32 crc kubenswrapper[4715]: I1203 21:43:32.634186 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:32 crc kubenswrapper[4715]: E1203 21:43:32.634288 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:32 crc kubenswrapper[4715]: I1203 21:43:32.634401 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:32 crc kubenswrapper[4715]: E1203 21:43:32.634540 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:32 crc kubenswrapper[4715]: E1203 21:43:32.634628 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:32 crc kubenswrapper[4715]: E1203 21:43:32.634816 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:34 crc kubenswrapper[4715]: I1203 21:43:34.633426 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:34 crc kubenswrapper[4715]: E1203 21:43:34.633699 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:34 crc kubenswrapper[4715]: I1203 21:43:34.633746 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:34 crc kubenswrapper[4715]: I1203 21:43:34.633862 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:34 crc kubenswrapper[4715]: E1203 21:43:34.634377 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:34 crc kubenswrapper[4715]: I1203 21:43:34.634802 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:34 crc kubenswrapper[4715]: E1203 21:43:34.634868 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:34 crc kubenswrapper[4715]: E1203 21:43:34.635104 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:36 crc kubenswrapper[4715]: I1203 21:43:36.633404 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:36 crc kubenswrapper[4715]: I1203 21:43:36.633476 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:36 crc kubenswrapper[4715]: I1203 21:43:36.633604 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:36 crc kubenswrapper[4715]: I1203 21:43:36.633667 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:36 crc kubenswrapper[4715]: E1203 21:43:36.633818 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:36 crc kubenswrapper[4715]: E1203 21:43:36.633991 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:36 crc kubenswrapper[4715]: E1203 21:43:36.634114 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:36 crc kubenswrapper[4715]: E1203 21:43:36.634247 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:38 crc kubenswrapper[4715]: I1203 21:43:38.636203 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:38 crc kubenswrapper[4715]: I1203 21:43:38.636268 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:38 crc kubenswrapper[4715]: I1203 21:43:38.636436 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:38 crc kubenswrapper[4715]: E1203 21:43:38.636557 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:38 crc kubenswrapper[4715]: I1203 21:43:38.636258 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:38 crc kubenswrapper[4715]: E1203 21:43:38.636659 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:38 crc kubenswrapper[4715]: E1203 21:43:38.636798 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:38 crc kubenswrapper[4715]: E1203 21:43:38.637066 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:40 crc kubenswrapper[4715]: I1203 21:43:40.636292 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:40 crc kubenswrapper[4715]: I1203 21:43:40.636307 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:40 crc kubenswrapper[4715]: I1203 21:43:40.636362 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:40 crc kubenswrapper[4715]: I1203 21:43:40.636360 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:40 crc kubenswrapper[4715]: E1203 21:43:40.636591 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:40 crc kubenswrapper[4715]: E1203 21:43:40.636717 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:40 crc kubenswrapper[4715]: E1203 21:43:40.636832 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:40 crc kubenswrapper[4715]: E1203 21:43:40.636904 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.357596 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/1.log" Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.358780 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/0.log" Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.358930 4715 generic.go:334] "Generic (PLEG): container finished" podID="9174d88f-f7f0-4b0c-942f-39d027b97e34" containerID="d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72" exitCode=1 Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.359104 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2wcn9" event={"ID":"9174d88f-f7f0-4b0c-942f-39d027b97e34","Type":"ContainerDied","Data":"d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72"} Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.359237 4715 scope.go:117] "RemoveContainer" containerID="0e12959b50be9aa90b872e7d4a1f11727464294d1fcb0e95a57dea938090c848" Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.360137 4715 scope.go:117] "RemoveContainer" containerID="d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72" Dec 03 21:43:42 crc kubenswrapper[4715]: E1203 21:43:42.360490 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2wcn9_openshift-multus(9174d88f-f7f0-4b0c-942f-39d027b97e34)\"" pod="openshift-multus/multus-2wcn9" podUID="9174d88f-f7f0-4b0c-942f-39d027b97e34" Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.634192 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.634263 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.634342 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:42 crc kubenswrapper[4715]: E1203 21:43:42.634487 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:42 crc kubenswrapper[4715]: I1203 21:43:42.634605 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:42 crc kubenswrapper[4715]: E1203 21:43:42.635159 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:42 crc kubenswrapper[4715]: E1203 21:43:42.635314 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:42 crc kubenswrapper[4715]: E1203 21:43:42.635546 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:43 crc kubenswrapper[4715]: I1203 21:43:43.365575 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/1.log" Dec 03 21:43:43 crc kubenswrapper[4715]: E1203 21:43:43.582017 4715 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 21:43:43 crc kubenswrapper[4715]: E1203 21:43:43.783592 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 21:43:44 crc kubenswrapper[4715]: I1203 21:43:44.633807 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:44 crc kubenswrapper[4715]: I1203 21:43:44.633931 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:44 crc kubenswrapper[4715]: E1203 21:43:44.634030 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:44 crc kubenswrapper[4715]: I1203 21:43:44.634104 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:44 crc kubenswrapper[4715]: E1203 21:43:44.634267 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:44 crc kubenswrapper[4715]: I1203 21:43:44.634344 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:44 crc kubenswrapper[4715]: E1203 21:43:44.634485 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:44 crc kubenswrapper[4715]: E1203 21:43:44.634631 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:44 crc kubenswrapper[4715]: I1203 21:43:44.635994 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 21:43:45 crc kubenswrapper[4715]: I1203 21:43:45.378591 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/3.log" Dec 03 21:43:45 crc kubenswrapper[4715]: I1203 21:43:45.382482 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerStarted","Data":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} Dec 03 21:43:45 crc kubenswrapper[4715]: I1203 21:43:45.382943 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:43:45 crc kubenswrapper[4715]: I1203 21:43:45.424732 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podStartSLOduration=97.424709317 podStartE2EDuration="1m37.424709317s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:43:45.424133072 +0000 UTC m=+122.166843697" watchObservedRunningTime="2025-12-03 21:43:45.424709317 +0000 UTC m=+122.167419952" Dec 03 21:43:45 crc kubenswrapper[4715]: I1203 21:43:45.690827 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hbsmx"] Dec 03 21:43:45 crc kubenswrapper[4715]: I1203 21:43:45.691051 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:45 crc kubenswrapper[4715]: E1203 21:43:45.691287 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:46 crc kubenswrapper[4715]: I1203 21:43:46.633999 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:46 crc kubenswrapper[4715]: I1203 21:43:46.634024 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:46 crc kubenswrapper[4715]: I1203 21:43:46.634024 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:46 crc kubenswrapper[4715]: E1203 21:43:46.634253 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:46 crc kubenswrapper[4715]: E1203 21:43:46.634456 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:46 crc kubenswrapper[4715]: E1203 21:43:46.634610 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:47 crc kubenswrapper[4715]: I1203 21:43:47.634325 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:47 crc kubenswrapper[4715]: E1203 21:43:47.634595 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:48 crc kubenswrapper[4715]: I1203 21:43:48.633838 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:48 crc kubenswrapper[4715]: E1203 21:43:48.634075 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:48 crc kubenswrapper[4715]: I1203 21:43:48.634444 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:48 crc kubenswrapper[4715]: E1203 21:43:48.634603 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:48 crc kubenswrapper[4715]: I1203 21:43:48.634898 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:48 crc kubenswrapper[4715]: E1203 21:43:48.635330 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:48 crc kubenswrapper[4715]: E1203 21:43:48.785742 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 21:43:49 crc kubenswrapper[4715]: I1203 21:43:49.634028 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:49 crc kubenswrapper[4715]: E1203 21:43:49.634294 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:50 crc kubenswrapper[4715]: I1203 21:43:50.633784 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:50 crc kubenswrapper[4715]: I1203 21:43:50.633805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:50 crc kubenswrapper[4715]: E1203 21:43:50.634711 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:50 crc kubenswrapper[4715]: I1203 21:43:50.633840 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:50 crc kubenswrapper[4715]: E1203 21:43:50.634713 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:50 crc kubenswrapper[4715]: E1203 21:43:50.635126 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:51 crc kubenswrapper[4715]: I1203 21:43:51.634150 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:51 crc kubenswrapper[4715]: E1203 21:43:51.634621 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:52 crc kubenswrapper[4715]: I1203 21:43:52.633669 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:52 crc kubenswrapper[4715]: I1203 21:43:52.633846 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:52 crc kubenswrapper[4715]: E1203 21:43:52.633914 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:52 crc kubenswrapper[4715]: I1203 21:43:52.633983 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:52 crc kubenswrapper[4715]: E1203 21:43:52.634196 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:52 crc kubenswrapper[4715]: E1203 21:43:52.634321 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:53 crc kubenswrapper[4715]: I1203 21:43:53.633994 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:53 crc kubenswrapper[4715]: E1203 21:43:53.636381 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:53 crc kubenswrapper[4715]: E1203 21:43:53.786786 4715 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 21:43:54 crc kubenswrapper[4715]: I1203 21:43:54.633366 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:54 crc kubenswrapper[4715]: I1203 21:43:54.633459 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:54 crc kubenswrapper[4715]: I1203 21:43:54.633570 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:54 crc kubenswrapper[4715]: E1203 21:43:54.633824 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:54 crc kubenswrapper[4715]: E1203 21:43:54.635206 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:54 crc kubenswrapper[4715]: E1203 21:43:54.635340 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:54 crc kubenswrapper[4715]: I1203 21:43:54.635698 4715 scope.go:117] "RemoveContainer" containerID="d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72" Dec 03 21:43:55 crc kubenswrapper[4715]: I1203 21:43:55.426988 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/1.log" Dec 03 21:43:55 crc kubenswrapper[4715]: I1203 21:43:55.427106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2wcn9" event={"ID":"9174d88f-f7f0-4b0c-942f-39d027b97e34","Type":"ContainerStarted","Data":"17360e95928ea8e902db39b173fee1daa0664875d571c8b1a6a2325810f6abbe"} Dec 03 21:43:55 crc kubenswrapper[4715]: I1203 21:43:55.633899 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:55 crc kubenswrapper[4715]: E1203 21:43:55.634102 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:56 crc kubenswrapper[4715]: I1203 21:43:56.634169 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:56 crc kubenswrapper[4715]: I1203 21:43:56.634339 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:56 crc kubenswrapper[4715]: I1203 21:43:56.635796 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:56 crc kubenswrapper[4715]: E1203 21:43:56.635991 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:56 crc kubenswrapper[4715]: E1203 21:43:56.636396 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:56 crc kubenswrapper[4715]: E1203 21:43:56.636495 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:57 crc kubenswrapper[4715]: I1203 21:43:57.634190 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:57 crc kubenswrapper[4715]: E1203 21:43:57.634460 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-hbsmx" podUID="34ccf04b-c534-43bb-94ee-fc2e7cef68c6" Dec 03 21:43:58 crc kubenswrapper[4715]: I1203 21:43:58.633422 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:43:58 crc kubenswrapper[4715]: I1203 21:43:58.633448 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:43:58 crc kubenswrapper[4715]: I1203 21:43:58.633587 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:43:58 crc kubenswrapper[4715]: E1203 21:43:58.633668 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 21:43:58 crc kubenswrapper[4715]: E1203 21:43:58.633860 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 21:43:58 crc kubenswrapper[4715]: E1203 21:43:58.633980 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 21:43:59 crc kubenswrapper[4715]: I1203 21:43:59.634137 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:43:59 crc kubenswrapper[4715]: I1203 21:43:59.637347 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 21:43:59 crc kubenswrapper[4715]: I1203 21:43:59.638160 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 21:44:00 crc kubenswrapper[4715]: I1203 21:44:00.633881 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:44:00 crc kubenswrapper[4715]: I1203 21:44:00.633970 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:44:00 crc kubenswrapper[4715]: I1203 21:44:00.634199 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:44:00 crc kubenswrapper[4715]: I1203 21:44:00.636622 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 21:44:00 crc kubenswrapper[4715]: I1203 21:44:00.638100 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 21:44:00 crc kubenswrapper[4715]: I1203 21:44:00.638555 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 21:44:00 crc kubenswrapper[4715]: I1203 21:44:00.638871 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 21:44:02 crc kubenswrapper[4715]: I1203 21:44:02.663997 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.082714 4715 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.138154 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bxzdz"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.139212 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.141959 4715 reflector.go:561] object-"openshift-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.142043 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.142529 4715 reflector.go:561] object-"openshift-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.142625 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.146396 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nmdf2"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.146896 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.147826 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-77n4d"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.148684 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.149985 4715 reflector.go:561] object-"openshift-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.150066 4715 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.150075 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.150133 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.150162 4715 reflector.go:561] object-"openshift-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.150222 4715 reflector.go:561] object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff": failed to list *v1.Secret: secrets "openshift-apiserver-sa-dockercfg-djjff" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.150225 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.150171 4715 reflector.go:561] object-"openshift-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.150300 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.150262 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-djjff\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-sa-dockercfg-djjff\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.150756 4715 reflector.go:561] object-"openshift-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.150790 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.152567 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.153955 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.154452 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.155063 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.156479 4715 reflector.go:561] object-"openshift-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.156568 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.156610 4715 reflector.go:561] object-"openshift-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.156624 4715 reflector.go:561] object-"openshift-apiserver"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.156652 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.156653 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.156670 4715 reflector.go:561] object-"openshift-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.156705 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164358 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164421 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpd6q\" (UniqueName: \"kubernetes.io/projected/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-kube-api-access-gpd6q\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164484 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164682 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-client\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164777 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b94r9\" (UniqueName: \"kubernetes.io/projected/311b832c-86ce-49c4-9455-daaca2e25480-kube-api-access-b94r9\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164813 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164839 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-client-ca\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164860 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-encryption-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.164977 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-images\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165070 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6mfb\" (UniqueName: \"kubernetes.io/projected/1ea0156d-4b30-4eb7-a285-a7becf5428ed-kube-api-access-v6mfb\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165121 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165165 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brxm7\" (UniqueName: \"kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165322 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-config\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165367 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-node-pullsecrets\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165427 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd7d035c-0bd3-4568-8384-03b463fb0c5c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165473 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165588 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgtbz\" (UniqueName: \"kubernetes.io/projected/cd7d035c-0bd3-4568-8384-03b463fb0c5c-kube-api-access-sgtbz\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165673 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165706 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165732 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/311b832c-86ce-49c4-9455-daaca2e25480-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165765 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165801 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit-dir\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165841 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-serving-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-config\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.165925 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ea0156d-4b30-4eb7-a285-a7becf5428ed-serving-cert\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.166033 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/311b832c-86ce-49c4-9455-daaca2e25480-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.171482 4715 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.171708 4715 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.171703 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.171771 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.173252 4715 reflector.go:561] object-"openshift-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.173353 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.173263 4715 reflector.go:561] object-"openshift-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.173419 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.173555 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.173565 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.173669 4715 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.173705 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.173852 4715 reflector.go:561] object-"openshift-controller-manager"/"openshift-global-ca": failed to list *v1.ConfigMap: configmaps "openshift-global-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.173890 4715 reflector.go:561] object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c": failed to list *v1.Secret: secrets "openshift-controller-manager-sa-dockercfg-msq4c" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.173949 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-msq4c\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-sa-dockercfg-msq4c\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.173894 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-global-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-global-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.174061 4715 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-images": failed to list *v1.ConfigMap: configmaps "machine-api-operator-images" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.174093 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-images\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-api-operator-images\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.174064 4715 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-tls": failed to list *v1.Secret: secrets "machine-api-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.174123 4715 reflector.go:561] object-"openshift-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.174144 4715 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.174198 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.174143 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.174141 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.174256 4715 reflector.go:561] object-"openshift-apiserver-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver-operator": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.174276 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.174398 4715 reflector.go:561] object-"openshift-machine-api"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.174425 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.174405 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.174791 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.175215 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.175214 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.175555 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 21:44:08 crc kubenswrapper[4715]: W1203 21:44:08.175773 4715 reflector.go:561] object-"openshift-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Dec 03 21:44:08 crc kubenswrapper[4715]: E1203 21:44:08.175806 4715 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.175825 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.176399 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.205680 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.206629 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.209645 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.210303 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rpjh4"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.210616 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.211044 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.211698 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.211926 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.212020 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.212781 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.217683 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.218274 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.218577 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.218629 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-fjvxw"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.218958 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.219048 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.219157 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.219230 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.219287 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-fjvxw" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.219336 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.219686 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.221399 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.221623 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.221744 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.221883 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222016 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222175 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222178 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qbjgj"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222300 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222407 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222529 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222605 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222833 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dzwxf"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.222994 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.223044 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.223211 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-sppb5"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.223222 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.223377 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.224625 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.224752 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.224784 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.226235 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.226380 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.227229 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.227232 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.228178 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.228620 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.228973 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.229002 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4w9hz"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.229720 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.235225 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z6hdq"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.235820 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.236353 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.236728 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.242794 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.248627 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.248870 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.249023 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.249207 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.249544 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.250397 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.250814 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.250934 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.251095 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.251211 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.251326 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.251371 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.251335 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.251599 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.251916 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252020 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252140 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252250 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252365 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252492 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252628 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252732 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252856 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.252955 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.253091 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.253198 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.253294 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.253395 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.253526 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.253645 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.254216 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-p4drr"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.274583 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.274932 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.277480 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.277696 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.278004 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.280303 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.285707 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300227 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300644 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpd6q\" (UniqueName: \"kubernetes.io/projected/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-kube-api-access-gpd6q\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300724 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ace1aa7b-9dc3-4dac-86f6-ddd994a3b554-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4w9hz\" (UID: \"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300757 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300779 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-client\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300795 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8565f32-5700-45f4-8176-3ba008f3b903-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300814 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b94r9\" (UniqueName: \"kubernetes.io/projected/311b832c-86ce-49c4-9455-daaca2e25480-kube-api-access-b94r9\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300836 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8565f32-5700-45f4-8176-3ba008f3b903-config\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-client-ca\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-encryption-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300911 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-images\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300930 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvx57\" (UniqueName: \"kubernetes.io/projected/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-kube-api-access-bvx57\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300948 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eaad1779-4e68-49e4-a2b3-7beebef05122-trusted-ca\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300966 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6mfb\" (UniqueName: \"kubernetes.io/projected/1ea0156d-4b30-4eb7-a285-a7becf5428ed-kube-api-access-v6mfb\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300983 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.300999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaad1779-4e68-49e4-a2b3-7beebef05122-config\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301015 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaad1779-4e68-49e4-a2b3-7beebef05122-serving-cert\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301034 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brxm7\" (UniqueName: \"kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301073 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-config\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301120 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301143 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-node-pullsecrets\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301166 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6zx6\" (UniqueName: \"kubernetes.io/projected/ace1aa7b-9dc3-4dac-86f6-ddd994a3b554-kube-api-access-g6zx6\") pod \"multus-admission-controller-857f4d67dd-4w9hz\" (UID: \"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301189 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd7d035c-0bd3-4568-8384-03b463fb0c5c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301192 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301225 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgtbz\" (UniqueName: \"kubernetes.io/projected/cd7d035c-0bd3-4568-8384-03b463fb0c5c-kube-api-access-sgtbz\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301242 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xgrw\" (UniqueName: \"kubernetes.io/projected/eaad1779-4e68-49e4-a2b3-7beebef05122-kube-api-access-4xgrw\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8565f32-5700-45f4-8176-3ba008f3b903-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301280 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301296 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301314 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301335 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/311b832c-86ce-49c4-9455-daaca2e25480-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301357 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301375 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301395 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit-dir\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301414 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-serving-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301430 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ea0156d-4b30-4eb7-a285-a7becf5428ed-serving-cert\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301446 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-config\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301469 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/311b832c-86ce-49c4-9455-daaca2e25480-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.302130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/311b832c-86ce-49c4-9455-daaca2e25480-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.303185 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-client-ca\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.301060 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.304143 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.304480 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-28k2p"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.304962 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.304945 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.305119 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit-dir\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.305257 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.306115 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.306217 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.306578 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.306711 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-config\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.308495 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.308547 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.308801 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.308928 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-node-pullsecrets\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.310350 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/311b832c-86ce-49c4-9455-daaca2e25480-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.310357 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.310466 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.310588 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.310991 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.311078 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.311442 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ea0156d-4b30-4eb7-a285-a7becf5428ed-serving-cert\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.311708 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.318858 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7n294"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.319618 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.320362 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.321309 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.321572 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.322174 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.322951 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.323493 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.324270 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.324805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.325123 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.326070 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.327388 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.327589 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.328032 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.328374 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cqrjp"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.328932 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.332620 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.333248 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.338724 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.340465 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pjddq"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.346915 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.347243 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.348757 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.351686 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.352637 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-77n4d"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.354466 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nmdf2"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.358248 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bxzdz"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.361107 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xrj4r"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.362451 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.366402 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.371866 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-vvbtt"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.373846 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vvbtt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.376311 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.382923 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.384222 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.385556 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.385910 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.387207 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qbjgj"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.388698 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sppb5"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.389555 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.391658 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-28k2p"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.391688 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.392907 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-fjvxw"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.394236 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.395142 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.396643 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rpjh4"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.397853 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.399035 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.401892 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ace1aa7b-9dc3-4dac-86f6-ddd994a3b554-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4w9hz\" (UID: \"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402287 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8565f32-5700-45f4-8176-3ba008f3b903-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402340 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8565f32-5700-45f4-8176-3ba008f3b903-config\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402376 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eaad1779-4e68-49e4-a2b3-7beebef05122-trusted-ca\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402402 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvx57\" (UniqueName: \"kubernetes.io/projected/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-kube-api-access-bvx57\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402435 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaad1779-4e68-49e4-a2b3-7beebef05122-config\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402455 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaad1779-4e68-49e4-a2b3-7beebef05122-serving-cert\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402522 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402553 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6zx6\" (UniqueName: \"kubernetes.io/projected/ace1aa7b-9dc3-4dac-86f6-ddd994a3b554-kube-api-access-g6zx6\") pod \"multus-admission-controller-857f4d67dd-4w9hz\" (UID: \"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402595 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xgrw\" (UniqueName: \"kubernetes.io/projected/eaad1779-4e68-49e4-a2b3-7beebef05122-kube-api-access-4xgrw\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402620 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8565f32-5700-45f4-8176-3ba008f3b903-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402644 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402668 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.402104 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4w9hz"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.404081 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8565f32-5700-45f4-8176-3ba008f3b903-config\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.404163 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.404709 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.404972 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eaad1779-4e68-49e4-a2b3-7beebef05122-config\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.405307 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.405932 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ace1aa7b-9dc3-4dac-86f6-ddd994a3b554-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4w9hz\" (UID: \"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.406224 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8565f32-5700-45f4-8176-3ba008f3b903-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.406937 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dzwxf"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.407258 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eaad1779-4e68-49e4-a2b3-7beebef05122-trusted-ca\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.407702 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.408355 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7n294"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.409241 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eaad1779-4e68-49e4-a2b3-7beebef05122-serving-cert\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.409672 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.410922 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.412131 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.413497 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.414761 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.415975 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vvbtt"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.417274 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.418500 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xrj4r"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.419672 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.420936 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.421811 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cqrjp"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.423045 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pjddq"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.424219 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z6hdq"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.425359 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.426447 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.427437 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-xtkkr"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.428781 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-lcvwz"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.428960 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.429658 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.429738 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.430746 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xtkkr"] Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.441963 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.461805 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.480890 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.588822 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6mfb\" (UniqueName: \"kubernetes.io/projected/1ea0156d-4b30-4eb7-a285-a7becf5428ed-kube-api-access-v6mfb\") pod \"route-controller-manager-6576b87f9c-48bw9\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.602169 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.621881 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.641616 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.661693 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.691953 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.701739 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.722156 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.743223 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.782053 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.802439 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.823184 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.843496 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.861984 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.884114 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.902701 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.922362 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.942497 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.962800 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 21:44:08 crc kubenswrapper[4715]: I1203 21:44:08.982093 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.002437 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.022614 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.042190 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.072328 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.097490 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.102967 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.121492 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.143181 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.162443 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.183033 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.202265 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.222553 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.242483 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.262124 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.284068 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.302710 4715 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.302808 4715 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.302844 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-client podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.802810172 +0000 UTC m=+146.545520797 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-client") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.302886 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-trusted-ca-bundle podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.802862363 +0000 UTC m=+146.545572998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-trusted-ca-bundle") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.302922 4715 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.302721 4715 secret.go:188] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.303061 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config podName:d4bdbca2-e0b6-4e6f-9749-fc76ba307440 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.803026727 +0000 UTC m=+146.545737332 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config") pod "controller-manager-879f6c89f-nmdf2" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.303096 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert podName:d4bdbca2-e0b6-4e6f-9749-fc76ba307440 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.803082689 +0000 UTC m=+146.545793314 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert") pod "controller-manager-879f6c89f-nmdf2" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440") : failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304242 4715 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304372 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-config podName:cd7d035c-0bd3-4568-8384-03b463fb0c5c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.804340513 +0000 UTC m=+146.547051148 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-config") pod "machine-api-operator-5694c8668f-77n4d" (UID: "cd7d035c-0bd3-4568-8384-03b463fb0c5c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304418 4715 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304461 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cd7d035c-0bd3-4568-8384-03b463fb0c5c-machine-api-operator-tls podName:cd7d035c-0bd3-4568-8384-03b463fb0c5c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.804448866 +0000 UTC m=+146.547159501 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/cd7d035c-0bd3-4568-8384-03b463fb0c5c-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-77n4d" (UID: "cd7d035c-0bd3-4568-8384-03b463fb0c5c") : failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304538 4715 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304577 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca podName:d4bdbca2-e0b6-4e6f-9749-fc76ba307440 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.804565869 +0000 UTC m=+146.547276504 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca") pod "controller-manager-879f6c89f-nmdf2" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304617 4715 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304641 4715 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304689 4715 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304705 4715 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304651 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-images podName:cd7d035c-0bd3-4568-8384-03b463fb0c5c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.804639751 +0000 UTC m=+146.547350386 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-images") pod "machine-api-operator-5694c8668f-77n4d" (UID: "cd7d035c-0bd3-4568-8384-03b463fb0c5c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304728 4715 secret.go:188] Couldn't get secret openshift-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304866 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.804809645 +0000 UTC m=+146.547520270 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304753 4715 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304987 4715 configmap.go:193] Couldn't get configMap openshift-controller-manager/openshift-global-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.305629 4715 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.304942 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.804893077 +0000 UTC m=+146.547603712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.306772 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.806725467 +0000 UTC m=+146.549436072 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.306803 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-encryption-config podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.806794319 +0000 UTC m=+146.549504934 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-encryption-config") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.306829 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.80682135 +0000 UTC m=+146.549531955 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.306847 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles podName:d4bdbca2-e0b6-4e6f-9749-fc76ba307440 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.80683796 +0000 UTC m=+146.549548565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-ca-bundles" (UniqueName: "kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles") pod "controller-manager-879f6c89f-nmdf2" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.306866 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-serving-ca podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:09.80685787 +0000 UTC m=+146.549568475 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-serving-ca") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.317871 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.322458 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.339735 4715 request.go:700] Waited for 1.017343049s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dcollect-profiles-dockercfg-kzf4t&limit=500&resourceVersion=0 Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.342112 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.361924 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.382600 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.403455 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.421688 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.442416 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.461944 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.481174 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.502668 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.513335 4715 projected.go:288] Couldn't get configMap openshift-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.521952 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.534437 4715 projected.go:288] Couldn't get configMap openshift-apiserver-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.534484 4715 projected.go:194] Error preparing data for projected volume kube-api-access-b94r9 for pod openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.534650 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/311b832c-86ce-49c4-9455-daaca2e25480-kube-api-access-b94r9 podName:311b832c-86ce-49c4-9455-daaca2e25480 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.034579508 +0000 UTC m=+146.777290143 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-b94r9" (UniqueName: "kubernetes.io/projected/311b832c-86ce-49c4-9455-daaca2e25480-kube-api-access-b94r9") pod "openshift-apiserver-operator-796bbdcf4f-5gnr6" (UID: "311b832c-86ce-49c4-9455-daaca2e25480") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.542313 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.558833 4715 projected.go:288] Couldn't get configMap openshift-machine-api/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.563133 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.581965 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.593144 4715 projected.go:288] Couldn't get configMap openshift-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.601684 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.605825 4715 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.605933 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.622580 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.642002 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.665326 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.683270 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.701397 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.722878 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.743262 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.763290 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.785761 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.803356 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.822474 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.823573 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-config\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.823664 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd7d035c-0bd3-4568-8384-03b463fb0c5c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.823706 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.823770 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.823813 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.823865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.823906 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-serving-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.823963 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.824006 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.824060 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.824103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-client\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.824188 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-encryption-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.824237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.824490 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-images\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.824564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.851038 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.861646 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.869083 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9"] Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.884875 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 21:44:09 crc kubenswrapper[4715]: W1203 21:44:09.885750 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ea0156d_4b30_4eb7_a285_a7becf5428ed.slice/crio-b749502ea9bd91bf7f40761f5198c20a7dedfb03684d9c95e899e41ecc74911b WatchSource:0}: Error finding container b749502ea9bd91bf7f40761f5198c20a7dedfb03684d9c95e899e41ecc74911b: Status 404 returned error can't find the container with id b749502ea9bd91bf7f40761f5198c20a7dedfb03684d9c95e899e41ecc74911b Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.927045 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.927461 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.927552 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-trusted-ca\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.927593 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-config\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.927642 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t79d7\" (UniqueName: \"kubernetes.io/projected/e439cfd1-5b9a-48d3-9789-0365a5d14d0f-kube-api-access-t79d7\") pod \"downloads-7954f5f757-fjvxw\" (UID: \"e439cfd1-5b9a-48d3-9789-0365a5d14d0f\") " pod="openshift-console/downloads-7954f5f757-fjvxw" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.927676 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-service-ca\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.927856 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40bf14c0-a69d-4de7-aef1-0233f8032d4c-audit-dir\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.927980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-oauth-serving-cert\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928089 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsvs4\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-kube-api-access-lsvs4\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928140 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-etcd-client\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928185 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3071aa31-a160-4695-9abe-9f94b838080f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928266 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/978ef8fd-0706-4780-b5fd-4cca61e3ea53-metrics-tls\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928375 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt8t2\" (UniqueName: \"kubernetes.io/projected/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-kube-api-access-pt8t2\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-tls\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928524 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-serving-cert\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7ppj\" (UniqueName: \"kubernetes.io/projected/fa7301e8-1cff-40a6-a83c-d8e2db119bce-kube-api-access-n7ppj\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928647 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928746 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-trusted-ca-bundle\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928789 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwsc6\" (UniqueName: \"kubernetes.io/projected/8988b24d-6be3-4293-8b20-fe20fab634e4-kube-api-access-qwsc6\") pod \"dns-operator-744455d44c-z6hdq\" (UID: \"8988b24d-6be3-4293-8b20-fe20fab634e4\") " pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhgmc\" (UniqueName: \"kubernetes.io/projected/3071aa31-a160-4695-9abe-9f94b838080f-kube-api-access-zhgmc\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928840 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928867 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928893 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/978ef8fd-0706-4780-b5fd-4cca61e3ea53-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.928918 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-encryption-config\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929129 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp8dd\" (UniqueName: \"kubernetes.io/projected/115e0818-142b-403c-8431-1aa9ced1d655-kube-api-access-mp8dd\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929292 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-serving-cert\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929391 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-oauth-config\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929441 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929593 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djnsk\" (UniqueName: \"kubernetes.io/projected/978ef8fd-0706-4780-b5fd-4cca61e3ea53-kube-api-access-djnsk\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929693 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929775 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929822 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-certificates\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929863 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.929980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-config\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/115e0818-142b-403c-8431-1aa9ced1d655-machine-approver-tls\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930221 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930266 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qbd4\" (UniqueName: \"kubernetes.io/projected/40bf14c0-a69d-4de7-aef1-0233f8032d4c-kube-api-access-9qbd4\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:09 crc kubenswrapper[4715]: E1203 21:44:09.930302 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.430285603 +0000 UTC m=+147.172996208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930334 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8988b24d-6be3-4293-8b20-fe20fab634e4-metrics-tls\") pod \"dns-operator-744455d44c-z6hdq\" (UID: \"8988b24d-6be3-4293-8b20-fe20fab634e4\") " pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930363 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930393 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-audit-policies\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930478 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3071aa31-a160-4695-9abe-9f94b838080f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930552 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/115e0818-142b-403c-8431-1aa9ced1d655-auth-proxy-config\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tfzb\" (UniqueName: \"kubernetes.io/projected/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-kube-api-access-9tfzb\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930854 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-service-ca-bundle\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.930949 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/978ef8fd-0706-4780-b5fd-4cca61e3ea53-trusted-ca\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.931015 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-config\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.931137 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-bound-sa-token\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.931259 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-serving-cert\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.931340 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/115e0818-142b-403c-8431-1aa9ced1d655-config\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.943611 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.963040 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 21:44:09 crc kubenswrapper[4715]: I1203 21:44:09.982088 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.003262 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.023131 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.032214 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.032394 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/978ef8fd-0706-4780-b5fd-4cca61e3ea53-metrics-tls\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.032474 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-tls\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.032558 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pt8t2\" (UniqueName: \"kubernetes.io/projected/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-kube-api-access-pt8t2\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.032641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zhwt\" (UniqueName: \"kubernetes.io/projected/4f726a35-ad6f-401c-92e6-de0a14416898-kube-api-access-9zhwt\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.032700 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.032789 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-registration-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.032957 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.532928216 +0000 UTC m=+147.275638821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.033219 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwsc6\" (UniqueName: \"kubernetes.io/projected/8988b24d-6be3-4293-8b20-fe20fab634e4-kube-api-access-qwsc6\") pod \"dns-operator-744455d44c-z6hdq\" (UID: \"8988b24d-6be3-4293-8b20-fe20fab634e4\") " pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.033486 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d7d78dd-f450-4538-9787-249b925548f4-proxy-tls\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.033597 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9ttk\" (UniqueName: \"kubernetes.io/projected/09c1b0b5-35b7-4090-893c-9f827e40527d-kube-api-access-x9ttk\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.033665 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-csi-data-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.033787 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/978ef8fd-0706-4780-b5fd-4cca61e3ea53-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.033855 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-metrics-certs\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.033914 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09c1b0b5-35b7-4090-893c-9f827e40527d-webhook-cert\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.033968 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034038 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1eba0929-2c1d-489f-a7c9-354c52b47b44-metrics-tls\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-serving-cert\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034167 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57v6h\" (UniqueName: \"kubernetes.io/projected/e39f2c0f-2634-4192-a66d-5f283f7f8e57-kube-api-access-57v6h\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034260 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djnsk\" (UniqueName: \"kubernetes.io/projected/978ef8fd-0706-4780-b5fd-4cca61e3ea53-kube-api-access-djnsk\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034341 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddef9904-1324-46d2-998f-0945fce96ad9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-brtc2\" (UID: \"ddef9904-1324-46d2-998f-0945fce96ad9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034402 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff41d8f3-cf63-47de-b67f-71f916df1e46-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034461 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-service-ca-bundle\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034596 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-certificates\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034676 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034744 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-config\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/115e0818-142b-403c-8431-1aa9ced1d655-machine-approver-tls\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034885 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-profile-collector-cert\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.034949 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-default-certificate\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.035021 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpvcg\" (UniqueName: \"kubernetes.io/projected/271da290-36a9-47c7-a055-15acac1e5b18-kube-api-access-xpvcg\") pod \"ingress-canary-vvbtt\" (UID: \"271da290-36a9-47c7-a055-15acac1e5b18\") " pod="openshift-ingress-canary/ingress-canary-vvbtt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.035089 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.035158 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.035224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qbd4\" (UniqueName: \"kubernetes.io/projected/40bf14c0-a69d-4de7-aef1-0233f8032d4c-kube-api-access-9qbd4\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.035492 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4p4s\" (UniqueName: \"kubernetes.io/projected/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-kube-api-access-z4p4s\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.035644 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2jk2\" (UniqueName: \"kubernetes.io/projected/20d62f2d-b30f-403f-aa04-c63e368599ae-kube-api-access-q2jk2\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.035730 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-service-ca\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.035841 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wghw\" (UniqueName: \"kubernetes.io/projected/ddef9904-1324-46d2-998f-0945fce96ad9-kube-api-access-9wghw\") pod \"package-server-manager-789f6589d5-brtc2\" (UID: \"ddef9904-1324-46d2-998f-0945fce96ad9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.036660 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.036249 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-config\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.036718 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66d4153f-10d5-4f92-8152-79509d22e17e-audit-dir\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.036905 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.036967 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcb4b\" (UniqueName: \"kubernetes.io/projected/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-kube-api-access-tcb4b\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.036996 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-config\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037053 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3071aa31-a160-4695-9abe-9f94b838080f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037075 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/115e0818-142b-403c-8431-1aa9ced1d655-auth-proxy-config\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-certificates\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037560 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km2qt\" (UniqueName: \"kubernetes.io/projected/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-kube-api-access-km2qt\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037650 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tfzb\" (UniqueName: \"kubernetes.io/projected/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-kube-api-access-9tfzb\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037696 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff41d8f3-cf63-47de-b67f-71f916df1e46-proxy-tls\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037738 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/115e0818-142b-403c-8431-1aa9ced1d655-auth-proxy-config\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037753 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037797 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g4nt\" (UniqueName: \"kubernetes.io/projected/ff41d8f3-cf63-47de-b67f-71f916df1e46-kube-api-access-8g4nt\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037839 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/978ef8fd-0706-4780-b5fd-4cca61e3ea53-trusted-ca\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037874 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-config\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037931 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp6vw\" (UniqueName: \"kubernetes.io/projected/1eba0929-2c1d-489f-a7c9-354c52b47b44-kube-api-access-zp6vw\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037966 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn6f6\" (UniqueName: \"kubernetes.io/projected/3d7d78dd-f450-4538-9787-249b925548f4-kube-api-access-kn6f6\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.037999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d88e13c-7334-49a9-822b-9ac54d6bf72f-secret-volume\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038030 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09c1b0b5-35b7-4090-893c-9f827e40527d-tmpfs\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038062 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09c1b0b5-35b7-4090-893c-9f827e40527d-apiservice-cert\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038096 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/271da290-36a9-47c7-a055-15acac1e5b18-cert\") pod \"ingress-canary-vvbtt\" (UID: \"271da290-36a9-47c7-a055-15acac1e5b18\") " pod="openshift-ingress-canary/ingress-canary-vvbtt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038130 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/115e0818-142b-403c-8431-1aa9ced1d655-config\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038161 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrpjs\" (UniqueName: \"kubernetes.io/projected/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-kube-api-access-qrpjs\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038194 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-node-bootstrap-token\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038225 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-trusted-ca\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038260 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-config\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038287 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038318 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t79d7\" (UniqueName: \"kubernetes.io/projected/e439cfd1-5b9a-48d3-9789-0365a5d14d0f-kube-api-access-t79d7\") pod \"downloads-7954f5f757-fjvxw\" (UID: \"e439cfd1-5b9a-48d3-9789-0365a5d14d0f\") " pod="openshift-console/downloads-7954f5f757-fjvxw" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038336 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038354 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn48x\" (UniqueName: \"kubernetes.io/projected/290986b1-4ac0-40df-9294-0cff8d471733-kube-api-access-fn48x\") pod \"control-plane-machine-set-operator-78cbb6b69f-dwjvx\" (UID: \"290986b1-4ac0-40df-9294-0cff8d471733\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038660 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-config\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038723 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-service-ca\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038761 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d7d78dd-f450-4538-9787-249b925548f4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038786 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p84xj\" (UniqueName: \"kubernetes.io/projected/e4f3359e-e96f-4a37-b8ca-880d94dffc59-kube-api-access-p84xj\") pod \"cluster-samples-operator-665b6dd947-bldms\" (UID: \"e4f3359e-e96f-4a37-b8ca-880d94dffc59\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.038809 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-certs\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039097 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-config\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039202 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-oauth-serving-cert\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039253 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsvs4\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-kube-api-access-lsvs4\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-etcd-client\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039324 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e4f3359e-e96f-4a37-b8ca-880d94dffc59-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bldms\" (UID: \"e4f3359e-e96f-4a37-b8ca-880d94dffc59\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039355 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a08d0a7d-2e87-4341-838d-a6571df91942-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039398 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-socket-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039418 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039443 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-mountpoint-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039479 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-serving-cert\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039488 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-service-ca\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039519 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7ppj\" (UniqueName: \"kubernetes.io/projected/fa7301e8-1cff-40a6-a83c-d8e2db119bce-kube-api-access-n7ppj\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039477 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/115e0818-142b-403c-8431-1aa9ced1d655-config\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039554 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b94r9\" (UniqueName: \"kubernetes.io/projected/311b832c-86ce-49c4-9455-daaca2e25480-kube-api-access-b94r9\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039579 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06fe03c8-83b2-42ce-85dc-20bc667bc58c-serving-cert\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039829 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.039907 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-srv-cert\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040040 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a08d0a7d-2e87-4341-838d-a6571df91942-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-trusted-ca\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040138 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-trusted-ca-bundle\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040202 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ff41d8f3-cf63-47de-b67f-71f916df1e46-images\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040265 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f726a35-ad6f-401c-92e6-de0a14416898-serving-cert\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040405 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhgmc\" (UniqueName: \"kubernetes.io/projected/3071aa31-a160-4695-9abe-9f94b838080f-kube-api-access-zhgmc\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040456 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040550 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040610 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040666 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040720 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-encryption-config\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.040775 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.041117 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-oauth-serving-cert\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.043219 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-ca-trust-extracted\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.043271 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.043771 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp8dd\" (UniqueName: \"kubernetes.io/projected/115e0818-142b-403c-8431-1aa9ced1d655-kube-api-access-mp8dd\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.042186 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-serving-cert\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.043848 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/978ef8fd-0706-4780-b5fd-4cca61e3ea53-metrics-tls\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.043970 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-tls\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.044021 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3071aa31-a160-4695-9abe-9f94b838080f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.044045 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n67nc\" (UniqueName: \"kubernetes.io/projected/1a9497d2-06a8-4c57-bb1b-a936019a9889-kube-api-access-n67nc\") pod \"migrator-59844c95c7-29fsw\" (UID: \"1a9497d2-06a8-4c57-bb1b-a936019a9889\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.044633 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/978ef8fd-0706-4780-b5fd-4cca61e3ea53-trusted-ca\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.044704 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.044710 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-oauth-config\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.045422 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.046063 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75pss\" (UniqueName: \"kubernetes.io/projected/06fe03c8-83b2-42ce-85dc-20bc667bc58c-kube-api-access-75pss\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.046309 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.046705 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eba0929-2c1d-489f-a7c9-354c52b47b44-config-volume\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.046950 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-trusted-ca-bundle\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.047161 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-srv-cert\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.048061 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-serving-cert\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.048098 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-etcd-client\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.048759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06fe03c8-83b2-42ce-85dc-20bc667bc58c-config\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.048911 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049017 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049110 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvldt\" (UniqueName: \"kubernetes.io/projected/a08d0a7d-2e87-4341-838d-a6571df91942-kube-api-access-gvldt\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049189 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8988b24d-6be3-4293-8b20-fe20fab634e4-metrics-tls\") pod \"dns-operator-744455d44c-z6hdq\" (UID: \"8988b24d-6be3-4293-8b20-fe20fab634e4\") " pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049273 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049373 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-audit-policies\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049454 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d88e13c-7334-49a9-822b-9ac54d6bf72f-config-volume\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049550 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.049653 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.549604843 +0000 UTC m=+147.292315468 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049740 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/20d62f2d-b30f-403f-aa04-c63e368599ae-signing-cabundle\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049811 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.049887 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050043 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-service-ca-bundle\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050108 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050147 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-ca\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050203 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050259 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050351 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-bound-sa-token\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050408 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn6mz\" (UniqueName: \"kubernetes.io/projected/66d4153f-10d5-4f92-8152-79509d22e17e-kube-api-access-rn6mz\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050459 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/290986b1-4ac0-40df-9294-0cff8d471733-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dwjvx\" (UID: \"290986b1-4ac0-40df-9294-0cff8d471733\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050540 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-stats-auth\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050706 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrtnn\" (UniqueName: \"kubernetes.io/projected/7d88e13c-7334-49a9-822b-9ac54d6bf72f-kube-api-access-lrtnn\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050767 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-serving-cert\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050829 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-plugins-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.050988 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051049 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-oauth-config\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051056 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgns6\" (UniqueName: \"kubernetes.io/projected/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-kube-api-access-mgns6\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051115 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-audit-policies\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051165 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/20d62f2d-b30f-403f-aa04-c63e368599ae-signing-key\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051250 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40bf14c0-a69d-4de7-aef1-0233f8032d4c-audit-dir\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051316 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-client\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3071aa31-a160-4695-9abe-9f94b838080f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051638 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-audit-policies\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.051048 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.052158 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-serving-cert\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.052951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-installation-pull-secrets\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.053306 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/40bf14c0-a69d-4de7-aef1-0233f8032d4c-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.053664 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3071aa31-a160-4695-9abe-9f94b838080f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.053853 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/40bf14c0-a69d-4de7-aef1-0233f8032d4c-audit-dir\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.054215 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-service-ca-bundle\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.056929 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-serving-cert\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.057937 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/8988b24d-6be3-4293-8b20-fe20fab634e4-metrics-tls\") pod \"dns-operator-744455d44c-z6hdq\" (UID: \"8988b24d-6be3-4293-8b20-fe20fab634e4\") " pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.060395 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/115e0818-142b-403c-8431-1aa9ced1d655-machine-approver-tls\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.062430 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/40bf14c0-a69d-4de7-aef1-0233f8032d4c-encryption-config\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.062948 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.083878 4715 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.103268 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.122178 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.142725 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.153454 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.153754 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.653705686 +0000 UTC m=+147.396416321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.153920 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e4f3359e-e96f-4a37-b8ca-880d94dffc59-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bldms\" (UID: \"e4f3359e-e96f-4a37-b8ca-880d94dffc59\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.153978 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a08d0a7d-2e87-4341-838d-a6571df91942-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154023 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-socket-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154103 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-mountpoint-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154169 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06fe03c8-83b2-42ce-85dc-20bc667bc58c-serving-cert\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-srv-cert\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154246 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a08d0a7d-2e87-4341-838d-a6571df91942-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154278 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-mountpoint-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154379 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ff41d8f3-cf63-47de-b67f-71f916df1e46-images\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154433 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f726a35-ad6f-401c-92e6-de0a14416898-serving-cert\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154483 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-socket-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.154497 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155310 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155314 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a08d0a7d-2e87-4341-838d-a6571df91942-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155374 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n67nc\" (UniqueName: \"kubernetes.io/projected/1a9497d2-06a8-4c57-bb1b-a936019a9889-kube-api-access-n67nc\") pod \"migrator-59844c95c7-29fsw\" (UID: \"1a9497d2-06a8-4c57-bb1b-a936019a9889\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155419 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75pss\" (UniqueName: \"kubernetes.io/projected/06fe03c8-83b2-42ce-85dc-20bc667bc58c-kube-api-access-75pss\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155451 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eba0929-2c1d-489f-a7c9-354c52b47b44-config-volume\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155483 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-srv-cert\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155531 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06fe03c8-83b2-42ce-85dc-20bc667bc58c-config\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155592 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155638 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvldt\" (UniqueName: \"kubernetes.io/projected/a08d0a7d-2e87-4341-838d-a6571df91942-kube-api-access-gvldt\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155687 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d88e13c-7334-49a9-822b-9ac54d6bf72f-config-volume\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155719 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155752 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/20d62f2d-b30f-403f-aa04-c63e368599ae-signing-cabundle\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155836 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155921 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155953 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-ca\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.155991 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156072 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn6mz\" (UniqueName: \"kubernetes.io/projected/66d4153f-10d5-4f92-8152-79509d22e17e-kube-api-access-rn6mz\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156101 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/290986b1-4ac0-40df-9294-0cff8d471733-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dwjvx\" (UID: \"290986b1-4ac0-40df-9294-0cff8d471733\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156135 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-stats-auth\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156174 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrtnn\" (UniqueName: \"kubernetes.io/projected/7d88e13c-7334-49a9-822b-9ac54d6bf72f-kube-api-access-lrtnn\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156206 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-plugins-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156245 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgns6\" (UniqueName: \"kubernetes.io/projected/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-kube-api-access-mgns6\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156273 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-audit-policies\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/20d62f2d-b30f-403f-aa04-c63e368599ae-signing-key\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156341 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-client\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156421 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zhwt\" (UniqueName: \"kubernetes.io/projected/4f726a35-ad6f-401c-92e6-de0a14416898-kube-api-access-9zhwt\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156451 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156492 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-registration-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d7d78dd-f450-4538-9787-249b925548f4-proxy-tls\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156624 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9ttk\" (UniqueName: \"kubernetes.io/projected/09c1b0b5-35b7-4090-893c-9f827e40527d-kube-api-access-x9ttk\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156650 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-csi-data-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-metrics-certs\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156709 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156727 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1eba0929-2c1d-489f-a7c9-354c52b47b44-metrics-tls\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156748 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09c1b0b5-35b7-4090-893c-9f827e40527d-webhook-cert\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156769 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57v6h\" (UniqueName: \"kubernetes.io/projected/e39f2c0f-2634-4192-a66d-5f283f7f8e57-kube-api-access-57v6h\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156821 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddef9904-1324-46d2-998f-0945fce96ad9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-brtc2\" (UID: \"ddef9904-1324-46d2-998f-0945fce96ad9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156850 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff41d8f3-cf63-47de-b67f-71f916df1e46-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156878 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-profile-collector-cert\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-default-certificate\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156929 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-service-ca-bundle\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156962 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpvcg\" (UniqueName: \"kubernetes.io/projected/271da290-36a9-47c7-a055-15acac1e5b18-kube-api-access-xpvcg\") pod \"ingress-canary-vvbtt\" (UID: \"271da290-36a9-47c7-a055-15acac1e5b18\") " pod="openshift-ingress-canary/ingress-canary-vvbtt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157000 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4p4s\" (UniqueName: \"kubernetes.io/projected/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-kube-api-access-z4p4s\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157024 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2jk2\" (UniqueName: \"kubernetes.io/projected/20d62f2d-b30f-403f-aa04-c63e368599ae-kube-api-access-q2jk2\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157052 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157102 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-service-ca\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157132 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66d4153f-10d5-4f92-8152-79509d22e17e-audit-dir\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157189 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcb4b\" (UniqueName: \"kubernetes.io/projected/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-kube-api-access-tcb4b\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157216 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-config\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157244 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wghw\" (UniqueName: \"kubernetes.io/projected/ddef9904-1324-46d2-998f-0945fce96ad9-kube-api-access-9wghw\") pod \"package-server-manager-789f6589d5-brtc2\" (UID: \"ddef9904-1324-46d2-998f-0945fce96ad9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157275 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km2qt\" (UniqueName: \"kubernetes.io/projected/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-kube-api-access-km2qt\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157303 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff41d8f3-cf63-47de-b67f-71f916df1e46-proxy-tls\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157331 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06fe03c8-83b2-42ce-85dc-20bc667bc58c-config\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157352 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157414 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g4nt\" (UniqueName: \"kubernetes.io/projected/ff41d8f3-cf63-47de-b67f-71f916df1e46-kube-api-access-8g4nt\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157487 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp6vw\" (UniqueName: \"kubernetes.io/projected/1eba0929-2c1d-489f-a7c9-354c52b47b44-kube-api-access-zp6vw\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157530 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d88e13c-7334-49a9-822b-9ac54d6bf72f-secret-volume\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157552 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09c1b0b5-35b7-4090-893c-9f827e40527d-tmpfs\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157573 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09c1b0b5-35b7-4090-893c-9f827e40527d-apiservice-cert\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157599 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/271da290-36a9-47c7-a055-15acac1e5b18-cert\") pod \"ingress-canary-vvbtt\" (UID: \"271da290-36a9-47c7-a055-15acac1e5b18\") " pod="openshift-ingress-canary/ingress-canary-vvbtt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn6f6\" (UniqueName: \"kubernetes.io/projected/3d7d78dd-f450-4538-9787-249b925548f4-kube-api-access-kn6f6\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157643 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrpjs\" (UniqueName: \"kubernetes.io/projected/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-kube-api-access-qrpjs\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157672 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-node-bootstrap-token\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157701 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157734 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn48x\" (UniqueName: \"kubernetes.io/projected/290986b1-4ac0-40df-9294-0cff8d471733-kube-api-access-fn48x\") pod \"control-plane-machine-set-operator-78cbb6b69f-dwjvx\" (UID: \"290986b1-4ac0-40df-9294-0cff8d471733\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157759 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d7d78dd-f450-4538-9787-249b925548f4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157784 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p84xj\" (UniqueName: \"kubernetes.io/projected/e4f3359e-e96f-4a37-b8ca-880d94dffc59-kube-api-access-p84xj\") pod \"cluster-samples-operator-665b6dd947-bldms\" (UID: \"e4f3359e-e96f-4a37-b8ca-880d94dffc59\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.157806 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-certs\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.158058 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.159184 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06fe03c8-83b2-42ce-85dc-20bc667bc58c-serving-cert\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.160071 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.660054886 +0000 UTC m=+147.402765481 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.160435 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3d7d78dd-f450-4538-9787-249b925548f4-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.160463 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.161192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d88e13c-7334-49a9-822b-9ac54d6bf72f-config-volume\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.161357 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a08d0a7d-2e87-4341-838d-a6571df91942-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.161563 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.161762 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-srv-cert\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.161790 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/e4f3359e-e96f-4a37-b8ca-880d94dffc59-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-bldms\" (UID: \"e4f3359e-e96f-4a37-b8ca-880d94dffc59\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.162055 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/20d62f2d-b30f-403f-aa04-c63e368599ae-signing-cabundle\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.162112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-plugins-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.163275 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.163754 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.163806 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.163850 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-audit-policies\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.164375 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-metrics-certs\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.164594 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-ca\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.165692 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.166129 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.167536 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/09c1b0b5-35b7-4090-893c-9f827e40527d-apiservice-cert\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.168465 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.168485 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.168664 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.169051 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3d7d78dd-f450-4538-9787-249b925548f4-proxy-tls\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.169051 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.156124 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ff41d8f3-cf63-47de-b67f-71f916df1e46-images\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.169339 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/290986b1-4ac0-40df-9294-0cff8d471733-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dwjvx\" (UID: \"290986b1-4ac0-40df-9294-0cff8d471733\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.169439 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f726a35-ad6f-401c-92e6-de0a14416898-serving-cert\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.169542 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66d4153f-10d5-4f92-8152-79509d22e17e-audit-dir\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.170497 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-service-ca\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.170626 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-csi-data-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.170693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f726a35-ad6f-401c-92e6-de0a14416898-config\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.171453 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.171628 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/20d62f2d-b30f-403f-aa04-c63e368599ae-signing-key\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.171884 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.172014 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e39f2c0f-2634-4192-a66d-5f283f7f8e57-registration-dir\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.172189 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.172287 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.172727 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-service-ca-bundle\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.173308 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff41d8f3-cf63-47de-b67f-71f916df1e46-auth-proxy-config\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.173388 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/09c1b0b5-35b7-4090-893c-9f827e40527d-tmpfs\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.174070 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-stats-auth\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.174357 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-srv-cert\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.174364 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ddef9904-1324-46d2-998f-0945fce96ad9-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-brtc2\" (UID: \"ddef9904-1324-46d2-998f-0945fce96ad9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.175211 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-default-certificate\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.176032 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/09c1b0b5-35b7-4090-893c-9f827e40527d-webhook-cert\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.176797 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4f726a35-ad6f-401c-92e6-de0a14416898-etcd-client\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.177648 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff41d8f3-cf63-47de-b67f-71f916df1e46-proxy-tls\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.178913 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d88e13c-7334-49a9-822b-9ac54d6bf72f-secret-volume\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.179689 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-profile-collector-cert\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.181150 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.181721 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.184061 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/271da290-36a9-47c7-a055-15acac1e5b18-cert\") pod \"ingress-canary-vvbtt\" (UID: \"271da290-36a9-47c7-a055-15acac1e5b18\") " pod="openshift-ingress-canary/ingress-canary-vvbtt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.231046 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a8565f32-5700-45f4-8176-3ba008f3b903-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lf7jz\" (UID: \"a8565f32-5700-45f4-8176-3ba008f3b903\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.232902 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.253093 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xgrw\" (UniqueName: \"kubernetes.io/projected/eaad1779-4e68-49e4-a2b3-7beebef05122-kube-api-access-4xgrw\") pod \"console-operator-58897d9998-dzwxf\" (UID: \"eaad1779-4e68-49e4-a2b3-7beebef05122\") " pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.259421 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.259962 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.759909755 +0000 UTC m=+147.502620520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.261210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.262004 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.7619778 +0000 UTC m=+147.504688405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.272056 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6zx6\" (UniqueName: \"kubernetes.io/projected/ace1aa7b-9dc3-4dac-86f6-ddd994a3b554-kube-api-access-g6zx6\") pod \"multus-admission-controller-857f4d67dd-4w9hz\" (UID: \"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.290717 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvx57\" (UniqueName: \"kubernetes.io/projected/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-kube-api-access-bvx57\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.302792 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.308083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1f162815-4fb6-48fc-a185-6a2d3b9b75e3-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-m4gx5\" (UID: \"1f162815-4fb6-48fc-a185-6a2d3b9b75e3\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.323134 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.337810 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/1eba0929-2c1d-489f-a7c9-354c52b47b44-metrics-tls\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.340580 4715 request.go:700] Waited for 1.910605738s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0 Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.342539 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.347373 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.348158 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eba0929-2c1d-489f-a7c9-354c52b47b44-config-volume\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.369208 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.370046 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.870018098 +0000 UTC m=+147.612728693 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.370145 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.370568 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.870560353 +0000 UTC m=+147.613270948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.371166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.407003 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.407791 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.413148 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-node-bootstrap-token\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.415189 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-certs\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.443061 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.444324 4715 projected.go:194] Error preparing data for projected volume kube-api-access-gpd6q for pod openshift-apiserver/apiserver-76f77b778f-bxzdz: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.444454 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-kube-api-access-gpd6q podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.944413363 +0000 UTC m=+147.687123978 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gpd6q" (UniqueName: "kubernetes.io/projected/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-kube-api-access-gpd6q") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.461247 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.463483 4715 projected.go:194] Error preparing data for projected volume kube-api-access-brxm7 for pod openshift-controller-manager/controller-manager-879f6c89f-nmdf2: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.463599 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7 podName:d4bdbca2-e0b6-4e6f-9749-fc76ba307440 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.963568247 +0000 UTC m=+147.706278852 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-brxm7" (UniqueName: "kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7") pod "controller-manager-879f6c89f-nmdf2" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.471464 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.472074 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.972022124 +0000 UTC m=+147.714732719 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.472248 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.472972 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:10.97296318 +0000 UTC m=+147.715673775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.493189 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.493439 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.496248 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.503952 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.507382 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" event={"ID":"1ea0156d-4b30-4eb7-a285-a7becf5428ed","Type":"ContainerStarted","Data":"37799e55bf1fed232c7710b7ced9fcd5e20a4d68e4b3467fe998d0364f0b898e"} Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.507441 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" event={"ID":"1ea0156d-4b30-4eb7-a285-a7becf5428ed","Type":"ContainerStarted","Data":"b749502ea9bd91bf7f40761f5198c20a7dedfb03684d9c95e899e41ecc74911b"} Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.507385 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.507867 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.518865 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.521401 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.535354 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz"] Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.543967 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 21:44:10 crc kubenswrapper[4715]: W1203 21:44:10.546868 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8565f32_5700_45f4_8176_3ba008f3b903.slice/crio-5ff2ebf1f83733005595f5d5101d7124ba5eb7895cb7769400dab8cb0a2eadba WatchSource:0}: Error finding container 5ff2ebf1f83733005595f5d5101d7124ba5eb7895cb7769400dab8cb0a2eadba: Status 404 returned error can't find the container with id 5ff2ebf1f83733005595f5d5101d7124ba5eb7895cb7769400dab8cb0a2eadba Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.550543 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.551286 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.559698 4715 projected.go:288] Couldn't get configMap openshift-machine-api/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.559735 4715 projected.go:194] Error preparing data for projected volume kube-api-access-sgtbz for pod openshift-machine-api/machine-api-operator-5694c8668f-77n4d: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.559821 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/cd7d035c-0bd3-4568-8384-03b463fb0c5c-kube-api-access-sgtbz podName:cd7d035c-0bd3-4568-8384-03b463fb0c5c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.059792519 +0000 UTC m=+147.802503114 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-sgtbz" (UniqueName: "kubernetes.io/projected/cd7d035c-0bd3-4568-8384-03b463fb0c5c-kube-api-access-sgtbz") pod "machine-api-operator-5694c8668f-77n4d" (UID: "cd7d035c-0bd3-4568-8384-03b463fb0c5c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.561824 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.571515 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-client\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.573527 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.573724 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.573888 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.574578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.574756 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.074716928 +0000 UTC m=+147.817427713 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.575165 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.575819 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.583633 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.584204 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.584596 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.585116 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.596831 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.605424 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.625072 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.626376 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-audit\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.627227 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5"] Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.643488 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.645814 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-config\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.662020 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.670578 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b94r9\" (UniqueName: \"kubernetes.io/projected/311b832c-86ce-49c4-9455-daaca2e25480-kube-api-access-b94r9\") pod \"openshift-apiserver-operator-796bbdcf4f-5gnr6\" (UID: \"311b832c-86ce-49c4-9455-daaca2e25480\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.676825 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.683259 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.688755 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.689423 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.189408165 +0000 UTC m=+147.932118760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.705092 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.714543 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-dzwxf"] Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.716274 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd7d035c-0bd3-4568-8384-03b463fb0c5c-images\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.721426 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 21:44:10 crc kubenswrapper[4715]: W1203 21:44:10.721811 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaad1779_4e68_49e4_a2b3_7beebef05122.slice/crio-e531e1efe377f1007680c05b9076afb79f7d0778031594b3db2322d64acc452a WatchSource:0}: Error finding container e531e1efe377f1007680c05b9076afb79f7d0778031594b3db2322d64acc452a: Status 404 returned error can't find the container with id e531e1efe377f1007680c05b9076afb79f7d0778031594b3db2322d64acc452a Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.733088 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/cd7d035c-0bd3-4568-8384-03b463fb0c5c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.744446 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.768658 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.775570 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.782452 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.789835 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.790491 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.290475306 +0000 UTC m=+148.033185901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.801060 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.806440 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-etcd-serving-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.821712 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.824213 4715 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.824297 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.824275593 +0000 UTC m=+148.566986188 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.824634 4715 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.824714 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.824691374 +0000 UTC m=+148.567401969 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.825362 4715 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.825402 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert podName:29f2ba8b-cbeb-4aba-b25f-1664dbdd744c nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.825394353 +0000 UTC m=+148.568104948 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert") pod "apiserver-76f77b778f-bxzdz" (UID: "29f2ba8b-cbeb-4aba-b25f-1664dbdd744c") : failed to sync secret cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.825472 4715 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.825581 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca podName:d4bdbca2-e0b6-4e6f-9749-fc76ba307440 nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.825556388 +0000 UTC m=+148.568266983 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca") pod "controller-manager-879f6c89f-nmdf2" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440") : failed to sync configmap cache: timed out waiting for the condition Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.828097 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-encryption-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:10 crc kubenswrapper[4715]: W1203 21:44:10.836602 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-0875dae3938d479d3e8f9a86464d47e19bc831a6812029d9d9a768f9f2c62f7d WatchSource:0}: Error finding container 0875dae3938d479d3e8f9a86464d47e19bc831a6812029d9d9a768f9f2c62f7d: Status 404 returned error can't find the container with id 0875dae3938d479d3e8f9a86464d47e19bc831a6812029d9d9a768f9f2c62f7d Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.841431 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.857175 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.861057 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.883910 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.883979 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.891615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.892020 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.391996599 +0000 UTC m=+148.134707194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.901414 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.909954 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6"] Dec 03 21:44:10 crc kubenswrapper[4715]: W1203 21:44:10.920571 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod311b832c_86ce_49c4_9455_daaca2e25480.slice/crio-0b44527a2d43a0e98d557b82103e69e483e329e81bf5259f0a75734b45706a0b WatchSource:0}: Error finding container 0b44527a2d43a0e98d557b82103e69e483e329e81bf5259f0a75734b45706a0b: Status 404 returned error can't find the container with id 0b44527a2d43a0e98d557b82103e69e483e329e81bf5259f0a75734b45706a0b Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.923257 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.942824 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.978211 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4w9hz"] Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.987267 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt8t2\" (UniqueName: \"kubernetes.io/projected/506c5e7a-6947-4a44-bb0b-2c06a66fbd7d-kube-api-access-pt8t2\") pod \"openshift-config-operator-7777fb866f-8gnrh\" (UID: \"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.993084 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.993360 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.493313697 +0000 UTC m=+148.236024292 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.993777 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpd6q\" (UniqueName: \"kubernetes.io/projected/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-kube-api-access-gpd6q\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.993868 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brxm7\" (UniqueName: \"kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:10 crc kubenswrapper[4715]: I1203 21:44:10.993939 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:10 crc kubenswrapper[4715]: E1203 21:44:10.994348 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.494341255 +0000 UTC m=+148.237051850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:10.998474 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brxm7\" (UniqueName: \"kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:10.998914 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpd6q\" (UniqueName: \"kubernetes.io/projected/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-kube-api-access-gpd6q\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.006723 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwsc6\" (UniqueName: \"kubernetes.io/projected/8988b24d-6be3-4293-8b20-fe20fab634e4-kube-api-access-qwsc6\") pod \"dns-operator-744455d44c-z6hdq\" (UID: \"8988b24d-6be3-4293-8b20-fe20fab634e4\") " pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.017308 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/978ef8fd-0706-4780-b5fd-4cca61e3ea53-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.040831 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djnsk\" (UniqueName: \"kubernetes.io/projected/978ef8fd-0706-4780-b5fd-4cca61e3ea53-kube-api-access-djnsk\") pod \"ingress-operator-5b745b69d9-sfxdk\" (UID: \"978ef8fd-0706-4780-b5fd-4cca61e3ea53\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.072773 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qbd4\" (UniqueName: \"kubernetes.io/projected/40bf14c0-a69d-4de7-aef1-0233f8032d4c-kube-api-access-9qbd4\") pod \"apiserver-7bbb656c7d-6vvkt\" (UID: \"40bf14c0-a69d-4de7-aef1-0233f8032d4c\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.079436 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tfzb\" (UniqueName: \"kubernetes.io/projected/1c7d35a8-3361-43ba-bf7f-66ab8c0d022e-kube-api-access-9tfzb\") pod \"authentication-operator-69f744f599-qbjgj\" (UID: \"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.094747 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.095245 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgtbz\" (UniqueName: \"kubernetes.io/projected/cd7d035c-0bd3-4568-8384-03b463fb0c5c-kube-api-access-sgtbz\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.098896 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.099181 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.599155416 +0000 UTC m=+148.341866011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.102322 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t79d7\" (UniqueName: \"kubernetes.io/projected/e439cfd1-5b9a-48d3-9789-0365a5d14d0f-kube-api-access-t79d7\") pod \"downloads-7954f5f757-fjvxw\" (UID: \"e439cfd1-5b9a-48d3-9789-0365a5d14d0f\") " pod="openshift-console/downloads-7954f5f757-fjvxw" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.115754 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgtbz\" (UniqueName: \"kubernetes.io/projected/cd7d035c-0bd3-4568-8384-03b463fb0c5c-kube-api-access-sgtbz\") pod \"machine-api-operator-5694c8668f-77n4d\" (UID: \"cd7d035c-0bd3-4568-8384-03b463fb0c5c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.124932 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.125204 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsvs4\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-kube-api-access-lsvs4\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.136973 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7ppj\" (UniqueName: \"kubernetes.io/projected/fa7301e8-1cff-40a6-a83c-d8e2db119bce-kube-api-access-n7ppj\") pod \"console-f9d7485db-sppb5\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.156363 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhgmc\" (UniqueName: \"kubernetes.io/projected/3071aa31-a160-4695-9abe-9f94b838080f-kube-api-access-zhgmc\") pod \"openshift-controller-manager-operator-756b6f6bc6-5phn2\" (UID: \"3071aa31-a160-4695-9abe-9f94b838080f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.159700 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.168492 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.186389 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp8dd\" (UniqueName: \"kubernetes.io/projected/115e0818-142b-403c-8431-1aa9ced1d655-kube-api-access-mp8dd\") pod \"machine-approver-56656f9798-kp77l\" (UID: \"115e0818-142b-403c-8431-1aa9ced1d655\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.197852 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.198386 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.698368068 +0000 UTC m=+148.441078673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.211107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-6pjlh\" (UID: \"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.215009 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-bound-sa-token\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.236046 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.245986 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n67nc\" (UniqueName: \"kubernetes.io/projected/1a9497d2-06a8-4c57-bb1b-a936019a9889-kube-api-access-n67nc\") pod \"migrator-59844c95c7-29fsw\" (UID: \"1a9497d2-06a8-4c57-bb1b-a936019a9889\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.260761 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn6f6\" (UniqueName: \"kubernetes.io/projected/3d7d78dd-f450-4538-9787-249b925548f4-kube-api-access-kn6f6\") pod \"machine-config-controller-84d6567774-m6mqx\" (UID: \"3d7d78dd-f450-4538-9787-249b925548f4\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.275187 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrpjs\" (UniqueName: \"kubernetes.io/projected/799b37a2-5dee-4d8d-954a-e7bd3f31bfde-kube-api-access-qrpjs\") pod \"router-default-5444994796-p4drr\" (UID: \"799b37a2-5dee-4d8d-954a-e7bd3f31bfde\") " pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.305418 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.306142 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.806115738 +0000 UTC m=+148.548826333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.309424 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvldt\" (UniqueName: \"kubernetes.io/projected/a08d0a7d-2e87-4341-838d-a6571df91942-kube-api-access-gvldt\") pod \"kube-storage-version-migrator-operator-b67b599dd-tnzdr\" (UID: \"a08d0a7d-2e87-4341-838d-a6571df91942\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.326930 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn48x\" (UniqueName: \"kubernetes.io/projected/290986b1-4ac0-40df-9294-0cff8d471733-kube-api-access-fn48x\") pod \"control-plane-machine-set-operator-78cbb6b69f-dwjvx\" (UID: \"290986b1-4ac0-40df-9294-0cff8d471733\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.331469 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.336095 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.336675 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.350162 4715 request.go:700] Waited for 1.188858709s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/serviceaccounts/collect-profiles/token Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.356876 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.369627 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qbjgj"] Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.372363 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-fjvxw" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.380954 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p84xj\" (UniqueName: \"kubernetes.io/projected/e4f3359e-e96f-4a37-b8ca-880d94dffc59-kube-api-access-p84xj\") pod \"cluster-samples-operator-665b6dd947-bldms\" (UID: \"e4f3359e-e96f-4a37-b8ca-880d94dffc59\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.382937 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrtnn\" (UniqueName: \"kubernetes.io/projected/7d88e13c-7334-49a9-822b-9ac54d6bf72f-kube-api-access-lrtnn\") pod \"collect-profiles-29413290-nwfsc\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.386177 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn6mz\" (UniqueName: \"kubernetes.io/projected/66d4153f-10d5-4f92-8152-79509d22e17e-kube-api-access-rn6mz\") pod \"oauth-openshift-558db77b4-7n294\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.404495 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.406897 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.407342 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:11.907328543 +0000 UTC m=+148.650039138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.426042 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zhwt\" (UniqueName: \"kubernetes.io/projected/4f726a35-ad6f-401c-92e6-de0a14416898-kube-api-access-9zhwt\") pod \"etcd-operator-b45778765-cqrjp\" (UID: \"4f726a35-ad6f-401c-92e6-de0a14416898\") " pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.442303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.442989 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgns6\" (UniqueName: \"kubernetes.io/projected/28a9cf4d-26e3-44b6-be9d-0d44f13571e2-kube-api-access-mgns6\") pod \"olm-operator-6b444d44fb-mphgr\" (UID: \"28a9cf4d-26e3-44b6-be9d-0d44f13571e2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.459130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c030ad5a-f4f3-422e-b7cc-5af2a4c515fe-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jhpz9\" (UID: \"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.476369 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.493723 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75pss\" (UniqueName: \"kubernetes.io/projected/06fe03c8-83b2-42ce-85dc-20bc667bc58c-kube-api-access-75pss\") pod \"service-ca-operator-777779d784-p7tgt\" (UID: \"06fe03c8-83b2-42ce-85dc-20bc667bc58c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.496398 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpvcg\" (UniqueName: \"kubernetes.io/projected/271da290-36a9-47c7-a055-15acac1e5b18-kube-api-access-xpvcg\") pod \"ingress-canary-vvbtt\" (UID: \"271da290-36a9-47c7-a055-15acac1e5b18\") " pod="openshift-ingress-canary/ingress-canary-vvbtt" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.503615 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.509641 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.509896 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.009856552 +0000 UTC m=+148.752567147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.514987 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" event={"ID":"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554","Type":"ContainerStarted","Data":"58d2cf930975eca8803a7cbb31a88bef3094fa19200bc926fe1eb1506673fd1b"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.515036 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" event={"ID":"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554","Type":"ContainerStarted","Data":"f0fe4ed74c7d98d133a91423edefa8a2d9c8d878408b7a768492fc0532adaba6"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.516955 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2jk2\" (UniqueName: \"kubernetes.io/projected/20d62f2d-b30f-403f-aa04-c63e368599ae-kube-api-access-q2jk2\") pod \"service-ca-9c57cc56f-pjddq\" (UID: \"20d62f2d-b30f-403f-aa04-c63e368599ae\") " pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.517566 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" event={"ID":"1f162815-4fb6-48fc-a185-6a2d3b9b75e3","Type":"ContainerStarted","Data":"6160ea4f6b71aafa0bf5744ff45debf546dc571dc631c1bfcc05feaf67639749"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.517623 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" event={"ID":"1f162815-4fb6-48fc-a185-6a2d3b9b75e3","Type":"ContainerStarted","Data":"995b013fae0fb92b869537baba0f94ec67697a58a36c7d3e431a7367936248db"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.519544 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" event={"ID":"a8565f32-5700-45f4-8176-3ba008f3b903","Type":"ContainerStarted","Data":"52e45d20bec46f13a7a72c0182148e9201dbc373924d0a2dbad4e001953dc70e"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.519691 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" event={"ID":"a8565f32-5700-45f4-8176-3ba008f3b903","Type":"ContainerStarted","Data":"5ff2ebf1f83733005595f5d5101d7124ba5eb7895cb7769400dab8cb0a2eadba"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.521444 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4p4s\" (UniqueName: \"kubernetes.io/projected/27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a-kube-api-access-z4p4s\") pod \"catalog-operator-68c6474976-8glbk\" (UID: \"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.522814 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.525113 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7941753f881a93d7f563ca54d2379b995167796c33516473777be27ada02e83b"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.525145 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"e30c97ac624d2194bc622789051ac067a868ff2a55118077afc854f4029b3930"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.526080 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.527152 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" event={"ID":"311b832c-86ce-49c4-9455-daaca2e25480","Type":"ContainerStarted","Data":"6bc625fcb98a11f0ab15ab64202b2a535a440e35c4f0bc859d3bbf791bd9247c"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.527185 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" event={"ID":"311b832c-86ce-49c4-9455-daaca2e25480","Type":"ContainerStarted","Data":"0b44527a2d43a0e98d557b82103e69e483e329e81bf5259f0a75734b45706a0b"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.530561 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.531390 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dzwxf" event={"ID":"eaad1779-4e68-49e4-a2b3-7beebef05122","Type":"ContainerStarted","Data":"f943d36a3734d053dadecec1e7f8f987fe4a369c6b6d5bd20d65401c680ccd88"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.531417 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-dzwxf" event={"ID":"eaad1779-4e68-49e4-a2b3-7beebef05122","Type":"ContainerStarted","Data":"e531e1efe377f1007680c05b9076afb79f7d0778031594b3db2322d64acc452a"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.532122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.534964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"566e3ea032f66f32f97c229780706a63d1fb6c8748bafab4621bf9aac3ea3cf5"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.534995 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"0875dae3938d479d3e8f9a86464d47e19bc831a6812029d9d9a768f9f2c62f7d"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.541745 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcb4b\" (UniqueName: \"kubernetes.io/projected/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-kube-api-access-tcb4b\") pod \"marketplace-operator-79b997595-28k2p\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.543331 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk"] Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.552323 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.552567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" event={"ID":"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e","Type":"ContainerStarted","Data":"3929f7786ebbe0ced01d5fa21d5e787b334fc0888dadd54638ff433d791d8d94"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.555710 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.562750 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.565319 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57v6h\" (UniqueName: \"kubernetes.io/projected/e39f2c0f-2634-4192-a66d-5f283f7f8e57-kube-api-access-57v6h\") pod \"csi-hostpathplugin-xrj4r\" (UID: \"e39f2c0f-2634-4192-a66d-5f283f7f8e57\") " pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.569930 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.577929 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"faef30d4a0e6cc839e851f7a23622f54a07903a4d303603e8805fe93a658aec3"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.577972 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"721a25aa698b38606c676a5754e53eccf917e31a5121824a2f09d4aaf0b71572"} Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.578260 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.579337 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wghw\" (UniqueName: \"kubernetes.io/projected/ddef9904-1324-46d2-998f-0945fce96ad9-kube-api-access-9wghw\") pod \"package-server-manager-789f6589d5-brtc2\" (UID: \"ddef9904-1324-46d2-998f-0945fce96ad9\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.589742 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.602894 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km2qt\" (UniqueName: \"kubernetes.io/projected/d718a5f5-9299-4f23-b1d4-d15e85cbda9a-kube-api-access-km2qt\") pod \"machine-config-server-lcvwz\" (UID: \"d718a5f5-9299-4f23-b1d4-d15e85cbda9a\") " pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.608736 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.612638 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.615492 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.115472846 +0000 UTC m=+148.858183441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.616284 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.627038 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp6vw\" (UniqueName: \"kubernetes.io/projected/1eba0929-2c1d-489f-a7c9-354c52b47b44-kube-api-access-zp6vw\") pod \"dns-default-xtkkr\" (UID: \"1eba0929-2c1d-489f-a7c9-354c52b47b44\") " pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.650049 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.653977 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g4nt\" (UniqueName: \"kubernetes.io/projected/ff41d8f3-cf63-47de-b67f-71f916df1e46-kube-api-access-8g4nt\") pod \"machine-config-operator-74547568cd-mn5lf\" (UID: \"ff41d8f3-cf63-47de-b67f-71f916df1e46\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.654279 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-vvbtt" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.662755 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9ttk\" (UniqueName: \"kubernetes.io/projected/09c1b0b5-35b7-4090-893c-9f827e40527d-kube-api-access-x9ttk\") pod \"packageserver-d55dfcdfc-69xnx\" (UID: \"09c1b0b5-35b7-4090-893c-9f827e40527d\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.664054 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.669750 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lcvwz" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.713433 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.714804 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.21478545 +0000 UTC m=+148.957496045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.740746 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-dzwxf" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.787050 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.794805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.813252 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.815868 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.816369 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.316352914 +0000 UTC m=+149.059063509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.845556 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.899968 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.901727 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh"] Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.917784 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.918133 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:11 crc kubenswrapper[4715]: E1203 21:44:11.918243 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.418211146 +0000 UTC m=+149.160921741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.918338 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.918380 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.918463 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.918954 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-config\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.919164 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca\") pod \"controller-manager-879f6c89f-nmdf2\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.929561 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-image-import-ca\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:11 crc kubenswrapper[4715]: I1203 21:44:11.948890 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/29f2ba8b-cbeb-4aba-b25f-1664dbdd744c-serving-cert\") pod \"apiserver-76f77b778f-bxzdz\" (UID: \"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c\") " pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.031583 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.031956 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.531943387 +0000 UTC m=+149.274653982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.071278 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.111644 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.133303 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.133990 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.633937293 +0000 UTC m=+149.376647888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.134057 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.134438 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.634422465 +0000 UTC m=+149.377133060 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.235653 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.236228 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.736201016 +0000 UTC m=+149.478911611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.318878 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-z6hdq"] Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.342009 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.344799 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.844774069 +0000 UTC m=+149.587484664 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.359240 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-77n4d"] Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.424356 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt"] Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.436707 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh"] Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.445605 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.445907 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:12.945889051 +0000 UTC m=+149.688599646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: W1203 21:44:12.526032 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8988b24d_6be3_4293_8b20_fe20fab634e4.slice/crio-c0901f3ce5c1530a808225d04c79c6e294603ee711a9d73b9a6b6188658a60f9 WatchSource:0}: Error finding container c0901f3ce5c1530a808225d04c79c6e294603ee711a9d73b9a6b6188658a60f9: Status 404 returned error can't find the container with id c0901f3ce5c1530a808225d04c79c6e294603ee711a9d73b9a6b6188658a60f9 Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.547326 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.547721 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.047705592 +0000 UTC m=+149.790416187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.587907 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" event={"ID":"cd7d035c-0bd3-4568-8384-03b463fb0c5c","Type":"ContainerStarted","Data":"d91faa13123cea986d115aaf6252892c31a48eb8dd6d82829a1a76e29ca0bf24"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.588936 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" event={"ID":"115e0818-142b-403c-8431-1aa9ced1d655","Type":"ContainerStarted","Data":"619021dbcc847733285fe9fc02deda5a88f16a827ca4c6ce54440d071816445b"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.588961 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" event={"ID":"115e0818-142b-403c-8431-1aa9ced1d655","Type":"ContainerStarted","Data":"84aaec6bb5e812d3542db981ec2177cd360eac694492f13a09f15abcc0aaffd2"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.590980 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lcvwz" event={"ID":"d718a5f5-9299-4f23-b1d4-d15e85cbda9a","Type":"ContainerStarted","Data":"0da428678aa708b0df14056d1edffea37482e4dcee2066920d38a7cb9ee1e085"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.591106 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lcvwz" event={"ID":"d718a5f5-9299-4f23-b1d4-d15e85cbda9a","Type":"ContainerStarted","Data":"dba39019306d4f98d22699179cb16a71c8cc812891bd11a3af3752fa42ce0faf"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.607392 4715 generic.go:334] "Generic (PLEG): container finished" podID="506c5e7a-6947-4a44-bb0b-2c06a66fbd7d" containerID="63b1d03638da01fa491efbfaa2aea9bd48f2410888a8a94dcdbf1ba5ca82705e" exitCode=0 Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.607480 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" event={"ID":"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d","Type":"ContainerDied","Data":"63b1d03638da01fa491efbfaa2aea9bd48f2410888a8a94dcdbf1ba5ca82705e"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.607540 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" event={"ID":"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d","Type":"ContainerStarted","Data":"2639c61cf5303d0543c16d330f0a5cc75dbf2caba99a3b3ccfac130a88ec1115"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.639850 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p4drr" event={"ID":"799b37a2-5dee-4d8d-954a-e7bd3f31bfde","Type":"ContainerStarted","Data":"d38a6cbb6c38671302f52b7498410330eda4d075036226ad62af1873c6200ad3"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.639896 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-p4drr" event={"ID":"799b37a2-5dee-4d8d-954a-e7bd3f31bfde","Type":"ContainerStarted","Data":"bd5a83f11d8668631a5fb39c04f8f54980f1e20b957de0c928753bd42b19a845"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.650357 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.657743 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" event={"ID":"1c7d35a8-3361-43ba-bf7f-66ab8c0d022e","Type":"ContainerStarted","Data":"4aceec151608d1eefd45f062d8f612be4f2be9418d4e710de0fc03bcb2f05c9a"} Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.659089 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.159049038 +0000 UTC m=+149.901759633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.659566 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.660190 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.160182319 +0000 UTC m=+149.902892914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.705125 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" event={"ID":"ace1aa7b-9dc3-4dac-86f6-ddd994a3b554","Type":"ContainerStarted","Data":"ab3f1993979017c13a99c62db6d600d34f3a2a0e267e4003f00d48b9b2fd345d"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.733497 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" event={"ID":"978ef8fd-0706-4780-b5fd-4cca61e3ea53","Type":"ContainerStarted","Data":"459ac7c4b7a6c051ad6846bd78e254f89857ae74d7ecf49e9914e718141a00bb"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.733598 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" event={"ID":"978ef8fd-0706-4780-b5fd-4cca61e3ea53","Type":"ContainerStarted","Data":"8d473558744cdf4f7f7a31d94b0e60f5dc60f50092b613d067a620fa487156a2"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.750867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" event={"ID":"8988b24d-6be3-4293-8b20-fe20fab634e4","Type":"ContainerStarted","Data":"c0901f3ce5c1530a808225d04c79c6e294603ee711a9d73b9a6b6188658a60f9"} Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.760702 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.763738 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.263717266 +0000 UTC m=+150.006427861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.841651 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-m4gx5" podStartSLOduration=124.841630155 podStartE2EDuration="2m4.841630155s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:12.815933106 +0000 UTC m=+149.558643701" watchObservedRunningTime="2025-12-03 21:44:12.841630155 +0000 UTC m=+149.584340750" Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.841870 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2"] Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.853982 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx"] Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.857122 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-fjvxw"] Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.862976 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.869217 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.369204955 +0000 UTC m=+150.111915550 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.967177 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:12 crc kubenswrapper[4715]: E1203 21:44:12.970064 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.47003874 +0000 UTC m=+150.212749335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:12 crc kubenswrapper[4715]: I1203 21:44:12.979908 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5gnr6" podStartSLOduration=124.979888824 podStartE2EDuration="2m4.979888824s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:12.979535315 +0000 UTC m=+149.722245910" watchObservedRunningTime="2025-12-03 21:44:12.979888824 +0000 UTC m=+149.722599419" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.051891 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-dzwxf" podStartSLOduration=125.051867665 podStartE2EDuration="2m5.051867665s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:13.050778446 +0000 UTC m=+149.793489041" watchObservedRunningTime="2025-12-03 21:44:13.051867665 +0000 UTC m=+149.794578250" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.071664 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.072055 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.572038146 +0000 UTC m=+150.314748741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.174317 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.174972 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.674954606 +0000 UTC m=+150.417665201 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.218111 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.233907 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.276540 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.279713 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.778800072 +0000 UTC m=+150.521510667 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.303254 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" podStartSLOduration=124.303210937 podStartE2EDuration="2m4.303210937s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:13.296668961 +0000 UTC m=+150.039379556" watchObservedRunningTime="2025-12-03 21:44:13.303210937 +0000 UTC m=+150.045921562" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.316401 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.319028 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.379574 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.379962 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.879943345 +0000 UTC m=+150.622653940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.413349 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lf7jz" podStartSLOduration=125.413328241 podStartE2EDuration="2m5.413328241s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:13.411882292 +0000 UTC m=+150.154592887" watchObservedRunningTime="2025-12-03 21:44:13.413328241 +0000 UTC m=+150.156038836" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.437733 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7n294"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.443604 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-vvbtt"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.478583 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.480969 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.481551 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:13.98153326 +0000 UTC m=+150.724243855 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.485732 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:13 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:13 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:13 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.485799 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.487810 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.496897 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-xtkkr"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.508556 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-sppb5"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.516026 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.517611 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-xrj4r"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.583625 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.584465 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.084322687 +0000 UTC m=+150.827033282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.584577 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:13 crc kubenswrapper[4715]: W1203 21:44:13.579647 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa7301e8_1cff_40a6_a83c_d8e2db119bce.slice/crio-4a5017defc8e37f5fe04667b690591659cd111f229e58b18bb90448a2de8e8c6 WatchSource:0}: Error finding container 4a5017defc8e37f5fe04667b690591659cd111f229e58b18bb90448a2de8e8c6: Status 404 returned error can't find the container with id 4a5017defc8e37f5fe04667b690591659cd111f229e58b18bb90448a2de8e8c6 Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.585763 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.085754016 +0000 UTC m=+150.828464611 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.632561 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.644083 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-28k2p"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.649237 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qbjgj" podStartSLOduration=125.649219758 podStartE2EDuration="2m5.649219758s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:13.648384626 +0000 UTC m=+150.391095221" watchObservedRunningTime="2025-12-03 21:44:13.649219758 +0000 UTC m=+150.391930343" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.691004 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.691819 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.191800621 +0000 UTC m=+150.934511216 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: W1203 21:44:13.691954 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff7576eb_0d55_4e08_9ee8_734dad0aa0d8.slice/crio-a4c15e9c44b258184e9f8eaab2661f0542ad09b29ced34d739a544cc560cde0e WatchSource:0}: Error finding container a4c15e9c44b258184e9f8eaab2661f0542ad09b29ced34d739a544cc560cde0e: Status 404 returned error can't find the container with id a4c15e9c44b258184e9f8eaab2661f0542ad09b29ced34d739a544cc560cde0e Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.726234 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.726284 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.726301 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.726313 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.736582 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-p4drr" podStartSLOduration=125.736556131 podStartE2EDuration="2m5.736556131s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:13.717244863 +0000 UTC m=+150.459955458" watchObservedRunningTime="2025-12-03 21:44:13.736556131 +0000 UTC m=+150.479266726" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.737309 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-cqrjp"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.749987 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nmdf2"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.750035 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-pjddq"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.784053 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bxzdz"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.796675 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.797040 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.297027293 +0000 UTC m=+151.039737888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.799919 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-4w9hz" podStartSLOduration=124.79989908 podStartE2EDuration="2m4.79989908s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:13.798014999 +0000 UTC m=+150.540725594" watchObservedRunningTime="2025-12-03 21:44:13.79989908 +0000 UTC m=+150.542609675" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.821804 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx"] Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.824586 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" event={"ID":"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8","Type":"ContainerStarted","Data":"a4c15e9c44b258184e9f8eaab2661f0542ad09b29ced34d739a544cc560cde0e"} Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.832259 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" event={"ID":"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a","Type":"ContainerStarted","Data":"86151c1ecec29f77a5804594eb8c789d4730b8be747122dd4d347385c7ef272c"} Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.835944 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" event={"ID":"cd7d035c-0bd3-4568-8384-03b463fb0c5c","Type":"ContainerStarted","Data":"372d4bf5aac9d8f8fe36a8177f992be8c8cdeb407c77b89a2f01dbf5c0c24941"} Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.835976 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" event={"ID":"cd7d035c-0bd3-4568-8384-03b463fb0c5c","Type":"ContainerStarted","Data":"ce3892494e22e2fb8cf4b26680d84bc72f327dbfb3d0aa019a568d579ff5da9a"} Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.862210 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-lcvwz" podStartSLOduration=5.86217333 podStartE2EDuration="5.86217333s" podCreationTimestamp="2025-12-03 21:44:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:13.848166204 +0000 UTC m=+150.590876799" watchObservedRunningTime="2025-12-03 21:44:13.86217333 +0000 UTC m=+150.604883935" Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.864467 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xtkkr" event={"ID":"1eba0929-2c1d-489f-a7c9-354c52b47b44","Type":"ContainerStarted","Data":"ed95ea01c5a5225f87ff1ee46024ad22576eb06030232721296bedd6ec8cc4a0"} Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.905470 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.905634 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.405607785 +0000 UTC m=+151.148318380 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.905891 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:13 crc kubenswrapper[4715]: E1203 21:44:13.907230 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.407222348 +0000 UTC m=+151.149932943 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:13 crc kubenswrapper[4715]: I1203 21:44:13.938915 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" event={"ID":"978ef8fd-0706-4780-b5fd-4cca61e3ea53","Type":"ContainerStarted","Data":"4272e7436555cadc3f6d2167b93c6eee52930ff8f4619efb1599a9e4fc996ebe"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.004814 4715 generic.go:334] "Generic (PLEG): container finished" podID="40bf14c0-a69d-4de7-aef1-0233f8032d4c" containerID="b9de7b298ef547e90323c0dd1a0c6d4b1d2b843c3c816ecfa7e5337743776a6f" exitCode=0 Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.004967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" event={"ID":"40bf14c0-a69d-4de7-aef1-0233f8032d4c","Type":"ContainerDied","Data":"b9de7b298ef547e90323c0dd1a0c6d4b1d2b843c3c816ecfa7e5337743776a6f"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.005015 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" event={"ID":"40bf14c0-a69d-4de7-aef1-0233f8032d4c","Type":"ContainerStarted","Data":"1d2b8832df0447097c812176d137ddf012dff62b03dd1ff1c5bdf168bbfdddb4"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.008522 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.014005 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.513966572 +0000 UTC m=+151.256677167 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.014800 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.017403 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.517380413 +0000 UTC m=+151.260091008 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.058080 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" event={"ID":"115e0818-142b-403c-8431-1aa9ced1d655","Type":"ContainerStarted","Data":"45c29bd2a95994b1912040df4cecff5f4f77090e5046cdae55b65925cbb9bd1f"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.101978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" event={"ID":"506c5e7a-6947-4a44-bb0b-2c06a66fbd7d","Type":"ContainerStarted","Data":"8a307bebccef0c653bfb669b852c838008f5396a204e572eba6b5a7be371369c"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.102414 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.121747 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" event={"ID":"28a9cf4d-26e3-44b6-be9d-0d44f13571e2","Type":"ContainerStarted","Data":"bbaa601f907925a1a38095d27201ca139cab20c68534aead0b63d14bd98c1b67"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.122696 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.122882 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.622857752 +0000 UTC m=+151.365568347 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.123175 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.124412 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.624388844 +0000 UTC m=+151.367099439 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.131835 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" event={"ID":"7d88e13c-7334-49a9-822b-9ac54d6bf72f","Type":"ContainerStarted","Data":"8ede262331fa124ea0cdc2b4cd88a12d50687d14d6c2892241eeb602d78016f3"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.134352 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sppb5" event={"ID":"fa7301e8-1cff-40a6-a83c-d8e2db119bce","Type":"ContainerStarted","Data":"4a5017defc8e37f5fe04667b690591659cd111f229e58b18bb90448a2de8e8c6"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.145535 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" event={"ID":"e39f2c0f-2634-4192-a66d-5f283f7f8e57","Type":"ContainerStarted","Data":"03ed45bc350b1bdf9550eeb4ab106228d5719f523c060fd4aa293d8d992a5716"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.187417 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" event={"ID":"3d7d78dd-f450-4538-9787-249b925548f4","Type":"ContainerStarted","Data":"0fc14ba72ebee1071efdcbff54b85553ccfa92b3178dcb217cc7d4cb6b0735b0"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.187881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" event={"ID":"3d7d78dd-f450-4538-9787-249b925548f4","Type":"ContainerStarted","Data":"5284d76a2e0ea05ebcc71fe67b695648653166165d7256d9fb065afe916d55a6"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.187892 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" event={"ID":"3d7d78dd-f450-4538-9787-249b925548f4","Type":"ContainerStarted","Data":"0ef1c3f161b329fd51c2b6f2f2dfc25e2a1c5275ffb1847e8fa39e19e89201c3"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.194753 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" event={"ID":"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe","Type":"ContainerStarted","Data":"7464f96e7fc40f623437969049cb5fd42d25d5fd869c559e21c16b5cb5c4b32f"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.219868 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" event={"ID":"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f","Type":"ContainerStarted","Data":"ffd42d91b4036292cd34f539adb7faad84208b510de8056438c852c32f3ee45b"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.219930 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" event={"ID":"4b63a90a-b2e9-497b-bc1b-cec6b1bbc98f","Type":"ContainerStarted","Data":"7c097b9cf91283798522585cc7ff8a5a0d5686751fda09490844449e14088b18"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.224185 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.224488 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.724464968 +0000 UTC m=+151.467175563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.224581 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.226145 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.726124803 +0000 UTC m=+151.468835398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.226825 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-fjvxw" event={"ID":"e439cfd1-5b9a-48d3-9789-0365a5d14d0f","Type":"ContainerStarted","Data":"129e3340744324b95d83ef22fee7356d1d15abc99d3b9cd72759b0dc53b13499"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.226850 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-fjvxw" event={"ID":"e439cfd1-5b9a-48d3-9789-0365a5d14d0f","Type":"ContainerStarted","Data":"a63f9ef590e5154ec753ed0545ed2f24f65ffb36ce8482d997a98e65d4c2a0a8"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.227822 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-fjvxw" Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.243165 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-fjvxw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.243403 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fjvxw" podUID="e439cfd1-5b9a-48d3-9789-0365a5d14d0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.248837 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" event={"ID":"3071aa31-a160-4695-9abe-9f94b838080f","Type":"ContainerStarted","Data":"9906a7175bb2c4a2d932b3415547707ced6a8d06d22a9b7616069321ec037b85"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.249160 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" event={"ID":"3071aa31-a160-4695-9abe-9f94b838080f","Type":"ContainerStarted","Data":"741702312a53be8d7ba1b27f8d0ba6ffc867050e3c61532de9a0a45cdeabf447"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.251390 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" event={"ID":"ff41d8f3-cf63-47de-b67f-71f916df1e46","Type":"ContainerStarted","Data":"7edce2f96b77455bd74d23ba72423b060cc61851768fe545d80242aaf2663da9"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.278064 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" event={"ID":"e4f3359e-e96f-4a37-b8ca-880d94dffc59","Type":"ContainerStarted","Data":"3dac120015abe50abb4b5f7c13397d4a3d1b763f1a0660e261cbc35baadfec31"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.286953 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vvbtt" event={"ID":"271da290-36a9-47c7-a055-15acac1e5b18","Type":"ContainerStarted","Data":"4859a3dbcf80583e5d185f667f09b847c5b8d9721a73dc696868fb7a8d8e6dcd"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.290537 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" event={"ID":"66d4153f-10d5-4f92-8152-79509d22e17e","Type":"ContainerStarted","Data":"165b93a19d20335daa4aa3b5097e07ee706b31b008e0ddd5bd5f47e11f475b63"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.293264 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" event={"ID":"06fe03c8-83b2-42ce-85dc-20bc667bc58c","Type":"ContainerStarted","Data":"e2de9cf01e9a866bc23b4e1cca9ac3bd8f6dd6df0e26638726e07b748c51e02c"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.296778 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" event={"ID":"1a9497d2-06a8-4c57-bb1b-a936019a9889","Type":"ContainerStarted","Data":"fd5402f0bf5a612082826a5c41d57a5592b172ac243cb3c675a9a83dde39f4b8"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.326052 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.326482 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.826464954 +0000 UTC m=+151.569175549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.330550 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" event={"ID":"8988b24d-6be3-4293-8b20-fe20fab634e4","Type":"ContainerStarted","Data":"fbbe8e0b181e3ea17793ef20dd8705fa8a8f32158c82cd76f5c6c9bd7ef71722"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.336436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" event={"ID":"09c1b0b5-35b7-4090-893c-9f827e40527d","Type":"ContainerStarted","Data":"3c4b56e8117f910f13104c1df24e4c4211aab1fe0e198e11ea6fa8868dfb3c2a"} Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.431979 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.434101 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:14.934085891 +0000 UTC m=+151.676796486 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.488649 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:14 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:14 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:14 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.488950 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.534467 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.534859 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.034839624 +0000 UTC m=+151.777550209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.636550 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.636999 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.136980644 +0000 UTC m=+151.879691239 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.687368 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-77n4d" podStartSLOduration=125.687347455 podStartE2EDuration="2m5.687347455s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:14.662012955 +0000 UTC m=+151.404723550" watchObservedRunningTime="2025-12-03 21:44:14.687347455 +0000 UTC m=+151.430058050" Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.739964 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.740494 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.24047611 +0000 UTC m=+151.983186705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.854719 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.855286 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.355270479 +0000 UTC m=+152.097981074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.940400 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kp77l" podStartSLOduration=126.940380561 podStartE2EDuration="2m6.940380561s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:14.900135343 +0000 UTC m=+151.642845938" watchObservedRunningTime="2025-12-03 21:44:14.940380561 +0000 UTC m=+151.683091156" Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.956428 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:14 crc kubenswrapper[4715]: E1203 21:44:14.956833 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.456817292 +0000 UTC m=+152.199527887 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.957093 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" podStartSLOduration=126.957072999 podStartE2EDuration="2m6.957072999s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:14.93844825 +0000 UTC m=+151.681158845" watchObservedRunningTime="2025-12-03 21:44:14.957072999 +0000 UTC m=+151.699783594" Dec 03 21:44:14 crc kubenswrapper[4715]: I1203 21:44:14.957953 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nmdf2"] Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.031369 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" podStartSLOduration=126.031348741 podStartE2EDuration="2m6.031348741s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:14.98690207 +0000 UTC m=+151.729612655" watchObservedRunningTime="2025-12-03 21:44:15.031348741 +0000 UTC m=+151.774059336" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.060292 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.060315 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sfxdk" podStartSLOduration=127.060290158 podStartE2EDuration="2m7.060290158s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.059623441 +0000 UTC m=+151.802334036" watchObservedRunningTime="2025-12-03 21:44:15.060290158 +0000 UTC m=+151.803000753" Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.060721 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.560705489 +0000 UTC m=+152.303416074 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.070554 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" podStartSLOduration=127.070534423 podStartE2EDuration="2m7.070534423s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.032378479 +0000 UTC m=+151.775089074" watchObservedRunningTime="2025-12-03 21:44:15.070534423 +0000 UTC m=+151.813245018" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.097786 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m6mqx" podStartSLOduration=126.097762163 podStartE2EDuration="2m6.097762163s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.096923151 +0000 UTC m=+151.839633746" watchObservedRunningTime="2025-12-03 21:44:15.097762163 +0000 UTC m=+151.840472758" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.146858 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-fjvxw" podStartSLOduration=127.14683786 podStartE2EDuration="2m7.14683786s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.144669811 +0000 UTC m=+151.887380406" watchObservedRunningTime="2025-12-03 21:44:15.14683786 +0000 UTC m=+151.889548455" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.168050 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.168221 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.668192372 +0000 UTC m=+152.410902967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.168267 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.168802 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.668793519 +0000 UTC m=+152.411504114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.186872 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" podStartSLOduration=127.186850773 podStartE2EDuration="2m7.186850773s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.185891798 +0000 UTC m=+151.928602393" watchObservedRunningTime="2025-12-03 21:44:15.186850773 +0000 UTC m=+151.929561368" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.225667 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5phn2" podStartSLOduration=127.225647604 podStartE2EDuration="2m7.225647604s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.220621719 +0000 UTC m=+151.963332314" watchObservedRunningTime="2025-12-03 21:44:15.225647604 +0000 UTC m=+151.968358189" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.269594 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.270049 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.770030344 +0000 UTC m=+152.512740939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.284021 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-6pjlh" podStartSLOduration=127.284001949 podStartE2EDuration="2m7.284001949s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.266934321 +0000 UTC m=+152.009644916" watchObservedRunningTime="2025-12-03 21:44:15.284001949 +0000 UTC m=+152.026712544" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.371077 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.371916 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.871900317 +0000 UTC m=+152.614610912 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.467314 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" event={"ID":"7d88e13c-7334-49a9-822b-9ac54d6bf72f","Type":"ContainerStarted","Data":"b8079d9db0601a1f5e4bc6ab874c873fc363425ea45c0846f5af709a2beb3626"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.469400 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" event={"ID":"ff41d8f3-cf63-47de-b67f-71f916df1e46","Type":"ContainerStarted","Data":"7ce2151a48afc8c8784e97829e1497bd44b9f4d7a4523b0211868077135b0dab"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.472375 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.473569 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:15.973550073 +0000 UTC m=+152.716260668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.486297 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" event={"ID":"20d62f2d-b30f-403f-aa04-c63e368599ae","Type":"ContainerStarted","Data":"01504f6336e1ae82608e133bf43a0c43c308fc733f975e9463bb8975e2d9dd68"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.492807 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:15 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:15 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:15 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.492857 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.502070 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" event={"ID":"66d4153f-10d5-4f92-8152-79509d22e17e","Type":"ContainerStarted","Data":"bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.503272 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.505876 4715 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-7n294 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" start-of-body= Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.505927 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" podUID="66d4153f-10d5-4f92-8152-79509d22e17e" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.515813 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" event={"ID":"27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a","Type":"ContainerStarted","Data":"da092b9ba24d987db49ec1d1821f17f055d8bc9e600a05669001956778336ffc"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.517287 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.519953 4715 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-8glbk container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.520022 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" podUID="27259a8a-cb55-4bfa-ae59-fff4d4bbdb9a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.520366 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" podStartSLOduration=126.520354609 podStartE2EDuration="2m6.520354609s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.51965717 +0000 UTC m=+152.262367765" watchObservedRunningTime="2025-12-03 21:44:15.520354609 +0000 UTC m=+152.263065204" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.537419 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" event={"ID":"e4f3359e-e96f-4a37-b8ca-880d94dffc59","Type":"ContainerStarted","Data":"1c1ca618e9a39c07cbb1ce7731a768ed76a3478bfdfb9f580b61f309ff5ac6f1"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.543403 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" event={"ID":"290986b1-4ac0-40df-9294-0cff8d471733","Type":"ContainerStarted","Data":"2c23ba80842ef1d804953fb3f67cad9f3e2cf3de095e28946a1d78961644f823"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.543462 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" event={"ID":"290986b1-4ac0-40df-9294-0cff8d471733","Type":"ContainerStarted","Data":"5ea0c8e741e0eba45e010bf1c0f525b09ac45344ce8017f05673cb6ae194f297"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.553904 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" podStartSLOduration=126.553883789 podStartE2EDuration="2m6.553883789s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.547020614 +0000 UTC m=+152.289731209" watchObservedRunningTime="2025-12-03 21:44:15.553883789 +0000 UTC m=+152.296594384" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.559760 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" event={"ID":"4f726a35-ad6f-401c-92e6-de0a14416898","Type":"ContainerStarted","Data":"5a88b985f021e6eb499bcc11a0f5746c881b100d1c64116efe517468c8b7a6f1"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.574219 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.578249 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.078236441 +0000 UTC m=+152.820947036 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.582269 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.585172 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.590813 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" event={"ID":"09c1b0b5-35b7-4090-893c-9f827e40527d","Type":"ContainerStarted","Data":"426aa2a2042b845a7fb9e00e8fdf68cfdc624df6fc9ad1064671cf40508fa9da"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.591799 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.592071 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.593349 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" podStartSLOduration=127.593330326 podStartE2EDuration="2m7.593330326s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.591999591 +0000 UTC m=+152.334710186" watchObservedRunningTime="2025-12-03 21:44:15.593330326 +0000 UTC m=+152.336040921" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.600429 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.628811 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.629305 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-vvbtt" event={"ID":"271da290-36a9-47c7-a055-15acac1e5b18","Type":"ContainerStarted","Data":"b26b8f2fbbda1ebb5b41e20ea85f99b1b50178d821d2285568002ad3bd8a5f89"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.675186 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.675709 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.175684776 +0000 UTC m=+152.918395371 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.691654 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" event={"ID":"d4bdbca2-e0b6-4e6f-9749-fc76ba307440","Type":"ContainerStarted","Data":"a283464962333e8b524f2825f690e035b24105123c194b06c8154747ff51af3e"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.691804 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" podUID="d4bdbca2-e0b6-4e6f-9749-fc76ba307440" containerName="controller-manager" containerID="cri-o://a4c14983b898ed988103d1d29c926152e2bae4a9c36a72197004e8d8acab466c" gracePeriod=30 Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.692222 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.704752 4715 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nmdf2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.704794 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" podUID="d4bdbca2-e0b6-4e6f-9749-fc76ba307440" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.739993 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" event={"ID":"1a9497d2-06a8-4c57-bb1b-a936019a9889","Type":"ContainerStarted","Data":"1e98dd9412312094a60f302caf34280a550cb24fa88cfb12d75265d0c027e6ff"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.756247 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dwjvx" podStartSLOduration=126.756226926 podStartE2EDuration="2m6.756226926s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.691348506 +0000 UTC m=+152.434059101" watchObservedRunningTime="2025-12-03 21:44:15.756226926 +0000 UTC m=+152.498937521" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.757910 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" podStartSLOduration=126.75790546 podStartE2EDuration="2m6.75790546s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.755031543 +0000 UTC m=+152.497742138" watchObservedRunningTime="2025-12-03 21:44:15.75790546 +0000 UTC m=+152.500616055" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.769295 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-z6hdq" event={"ID":"8988b24d-6be3-4293-8b20-fe20fab634e4","Type":"ContainerStarted","Data":"d6823a23fdfca0e0682b67fcaaa5cce8ce8619320e2dd214e821938b26dec313"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.773138 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" event={"ID":"c030ad5a-f4f3-422e-b7cc-5af2a4c515fe","Type":"ContainerStarted","Data":"e941abc13c5285fae4bd492a0f59f35079c9d653a635bfd44cf252ed9d98d1e0"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.776827 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.777251 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2977a90b-71d1-450a-9551-0d375dc5d691-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2977a90b-71d1-450a-9551-0d375dc5d691\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.777315 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2977a90b-71d1-450a-9551-0d375dc5d691-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2977a90b-71d1-450a-9551-0d375dc5d691\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.777885 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.277869187 +0000 UTC m=+153.020579782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.828315 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-vvbtt" podStartSLOduration=7.828292289 podStartE2EDuration="7.828292289s" podCreationTimestamp="2025-12-03 21:44:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.79964707 +0000 UTC m=+152.542357665" watchObservedRunningTime="2025-12-03 21:44:15.828292289 +0000 UTC m=+152.571002884" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.830945 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sppb5" event={"ID":"fa7301e8-1cff-40a6-a83c-d8e2db119bce","Type":"ContainerStarted","Data":"5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.861473 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jhpz9" podStartSLOduration=127.861452468 podStartE2EDuration="2m7.861452468s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.859822114 +0000 UTC m=+152.602532709" watchObservedRunningTime="2025-12-03 21:44:15.861452468 +0000 UTC m=+152.604163063" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.862119 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" podStartSLOduration=127.862112126 podStartE2EDuration="2m7.862112126s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.830783725 +0000 UTC m=+152.573494320" watchObservedRunningTime="2025-12-03 21:44:15.862112126 +0000 UTC m=+152.604822721" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.870092 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" event={"ID":"a08d0a7d-2e87-4341-838d-a6571df91942","Type":"ContainerStarted","Data":"5a4c521615c6044b268dfa176879bb9ae75c02add5726696c5fa4316dcaa494b"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.870158 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" event={"ID":"a08d0a7d-2e87-4341-838d-a6571df91942","Type":"ContainerStarted","Data":"8116a1b6b984b3b87dc8b68b0af587f8877ae73946cf25ef645d3d79a71c8400"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.878102 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.878204 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.378172986 +0000 UTC m=+153.120883581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.879067 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.879159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2977a90b-71d1-450a-9551-0d375dc5d691-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2977a90b-71d1-450a-9551-0d375dc5d691\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.879218 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2977a90b-71d1-450a-9551-0d375dc5d691-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2977a90b-71d1-450a-9551-0d375dc5d691\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:15 crc kubenswrapper[4715]: E1203 21:44:15.881352 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.381334481 +0000 UTC m=+153.124045076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.881474 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2977a90b-71d1-450a-9551-0d375dc5d691-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2977a90b-71d1-450a-9551-0d375dc5d691\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.903585 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-sppb5" podStartSLOduration=127.903560547 podStartE2EDuration="2m7.903560547s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.900258739 +0000 UTC m=+152.642969334" watchObservedRunningTime="2025-12-03 21:44:15.903560547 +0000 UTC m=+152.646271132" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.914802 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-p7tgt" event={"ID":"06fe03c8-83b2-42ce-85dc-20bc667bc58c","Type":"ContainerStarted","Data":"c6ece4bbcee9430e540eb5022ff330b8374d3a66aaf57ef43c4d3f5ce1ce9135"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.930072 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2977a90b-71d1-450a-9551-0d375dc5d691-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2977a90b-71d1-450a-9551-0d375dc5d691\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.933303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.936097 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tnzdr" podStartSLOduration=127.93607745 podStartE2EDuration="2m7.93607745s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:15.933198782 +0000 UTC m=+152.675909377" watchObservedRunningTime="2025-12-03 21:44:15.93607745 +0000 UTC m=+152.678788045" Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.969452 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" event={"ID":"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8","Type":"ContainerStarted","Data":"fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab"} Dec 03 21:44:15 crc kubenswrapper[4715]: I1203 21:44:15.971210 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:15.998283 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-28k2p container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:15.998325 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:15.998337 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:15.999226 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.499166663 +0000 UTC m=+153.241877258 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.019242 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" event={"ID":"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c","Type":"ContainerStarted","Data":"0fcacf5d0eea3bc3acab7c5dc032e9702451180d52bd79c019e862eee8b10204"} Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.031077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" event={"ID":"ddef9904-1324-46d2-998f-0945fce96ad9","Type":"ContainerStarted","Data":"cdd965b7577a5ebbd2c0f228d2baa81568667a3a3011b0645198982ecfe4c090"} Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.031147 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" event={"ID":"ddef9904-1324-46d2-998f-0945fce96ad9","Type":"ContainerStarted","Data":"a6f3e037cd11106ef6865cdd41b682e16d520a4a21f3df0005c7943420097264"} Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.032792 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.059306 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" podStartSLOduration=127.059288285 podStartE2EDuration="2m7.059288285s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:16.024630685 +0000 UTC m=+152.767341270" watchObservedRunningTime="2025-12-03 21:44:16.059288285 +0000 UTC m=+152.801998880" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.094432 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" podStartSLOduration=127.094411957 podStartE2EDuration="2m7.094411957s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:16.094028256 +0000 UTC m=+152.836738851" watchObservedRunningTime="2025-12-03 21:44:16.094411957 +0000 UTC m=+152.837122552" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.113647 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.123585 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.623556028 +0000 UTC m=+153.366266613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.125950 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xtkkr" event={"ID":"1eba0929-2c1d-489f-a7c9-354c52b47b44","Type":"ContainerStarted","Data":"93908a9074bddeb33a275fb7bc0fc59deb7e017c5c98d54bb9076072a3672b66"} Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.127336 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-fjvxw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.127381 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fjvxw" podUID="e439cfd1-5b9a-48d3-9789-0365a5d14d0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.135858 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-8gnrh" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.170903 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-xtkkr" podStartSLOduration=8.170876938 podStartE2EDuration="8.170876938s" podCreationTimestamp="2025-12-03 21:44:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:16.16424738 +0000 UTC m=+152.906957995" watchObservedRunningTime="2025-12-03 21:44:16.170876938 +0000 UTC m=+152.913587533" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.220628 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.221287 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.721264429 +0000 UTC m=+153.463975024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.221518 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.226396 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.726375417 +0000 UTC m=+153.469086012 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.325185 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.325688 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.82566381 +0000 UTC m=+153.568374405 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.402544 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-69xnx" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.429185 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.429672 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:16.929659089 +0000 UTC m=+153.672369684 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.482915 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:16 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:16 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:16 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.482977 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.530923 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.531308 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.031271686 +0000 UTC m=+153.773982281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.531620 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.532092 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.032077197 +0000 UTC m=+153.774787792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.543698 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.633087 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.634699 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.134664168 +0000 UTC m=+153.877374763 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.737661 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.738169 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.238153794 +0000 UTC m=+153.980864389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.839485 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.840379 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.340342305 +0000 UTC m=+154.083052900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:16 crc kubenswrapper[4715]: I1203 21:44:16.941471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:16 crc kubenswrapper[4715]: E1203 21:44:16.941862 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.441843298 +0000 UTC m=+154.184553893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.042555 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.042794 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.542772505 +0000 UTC m=+154.285483100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.042865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.043324 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.543303409 +0000 UTC m=+154.286014174 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.121347 4715 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.137734 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" event={"ID":"d4bdbca2-e0b6-4e6f-9749-fc76ba307440","Type":"ContainerStarted","Data":"a4c14983b898ed988103d1d29c926152e2bae4a9c36a72197004e8d8acab466c"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.143599 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-xtkkr" event={"ID":"1eba0929-2c1d-489f-a7c9-354c52b47b44","Type":"ContainerStarted","Data":"ba92b92402c9ee06fe396c214a4d13bb7955ff9c8e01aaba3e0de9a0e733925f"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.143673 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.143882 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.643856687 +0000 UTC m=+154.386567282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.143903 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.144148 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.144629 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.644613557 +0000 UTC m=+154.387324152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.149558 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-pjddq" event={"ID":"20d62f2d-b30f-403f-aa04-c63e368599ae","Type":"ContainerStarted","Data":"8d301a72f9014153e456e1dbd32a19f4549e34d943ddbb794c2a4263cf6ae8e3"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.153955 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" event={"ID":"28a9cf4d-26e3-44b6-be9d-0d44f13571e2","Type":"ContainerStarted","Data":"ff009b807440f1a212e6677f321af2e0aaa2554f9031b9140e9f51e650c141bd"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.154179 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.156218 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" event={"ID":"e4f3359e-e96f-4a37-b8ca-880d94dffc59","Type":"ContainerStarted","Data":"665f31bbc915d0d031929945392a4cfdc5743e36e61ae32683a8ef6637caf761"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.158331 4715 generic.go:334] "Generic (PLEG): container finished" podID="29f2ba8b-cbeb-4aba-b25f-1664dbdd744c" containerID="6b28e6a667eaeb7044f3d048e4ee2097986e16991d7acbae0a1dd313735119c0" exitCode=0 Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.158411 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" event={"ID":"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c","Type":"ContainerStarted","Data":"a1fd98410423ff054d179cac2b51b333d54f0f3383a5b56993b0ea0ab781087f"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.158455 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" event={"ID":"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c","Type":"ContainerStarted","Data":"156346020bb98027c3bac74298af9b0c5c418b4f853b091bc9e2ebbf5e26079b"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.158468 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" event={"ID":"29f2ba8b-cbeb-4aba-b25f-1664dbdd744c","Type":"ContainerDied","Data":"6b28e6a667eaeb7044f3d048e4ee2097986e16991d7acbae0a1dd313735119c0"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.161389 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" event={"ID":"ddef9904-1324-46d2-998f-0945fce96ad9","Type":"ContainerStarted","Data":"1ba15d84abdbc7272fd4a7c212f230d956b9eed6be2abeb4a4f2769f89b6ecce"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.168555 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2977a90b-71d1-450a-9551-0d375dc5d691","Type":"ContainerStarted","Data":"24c38ed5921f996fc2ff25840b8f3a923d76a809a3d3f41526dc3c3fe27c8fa7"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.168602 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2977a90b-71d1-450a-9551-0d375dc5d691","Type":"ContainerStarted","Data":"fad1aad178196b92b00f02d96c5afa8baef59b99752e299735d3080a21ee944d"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.175048 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" event={"ID":"ff41d8f3-cf63-47de-b67f-71f916df1e46","Type":"ContainerStarted","Data":"f115be730f987a1d0a93d582ad069d2e67487c9c271691874ac4ddf4e17af620"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.177765 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" event={"ID":"e39f2c0f-2634-4192-a66d-5f283f7f8e57","Type":"ContainerStarted","Data":"9c5d73fbaaf36f9e4f4c2962c9a8415244c77e5586b6670dd7f827592b20efbb"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.177792 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" event={"ID":"e39f2c0f-2634-4192-a66d-5f283f7f8e57","Type":"ContainerStarted","Data":"61aa167b713e75e66b117bef1613bcf18eef7764c28fd3b3d6bb0d037ad801d7"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.178985 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" event={"ID":"1a9497d2-06a8-4c57-bb1b-a936019a9889","Type":"ContainerStarted","Data":"afa6a31a81118f99b971e845f14963ef4a6cea9bbbd011491761e8afb543bd3a"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.180383 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" event={"ID":"4f726a35-ad6f-401c-92e6-de0a14416898","Type":"ContainerStarted","Data":"777ef9be24f589d309fa76bbd386b84323258e00d41d2c7339880e5b6f0e04f4"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.181486 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.183610 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" event={"ID":"40bf14c0-a69d-4de7-aef1-0233f8032d4c","Type":"ContainerStarted","Data":"c80975b3039acf49da0c1e5eedf0a56096367ac33a69bc90349a2f2525e73430"} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.185530 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-fjvxw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.185596 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fjvxw" podUID="e439cfd1-5b9a-48d3-9789-0365a5d14d0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.186538 4715 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-28k2p container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.186581 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.17:8080/healthz\": dial tcp 10.217.0.17:8080: connect: connection refused" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.195124 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.214887 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8glbk" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.228237 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mphgr" podStartSLOduration=128.228208679 podStartE2EDuration="2m8.228208679s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:17.202362076 +0000 UTC m=+153.945072671" watchObservedRunningTime="2025-12-03 21:44:17.228208679 +0000 UTC m=+153.970919274" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.245291 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-cqrjp" podStartSLOduration=129.245271238 podStartE2EDuration="2m9.245271238s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:17.242585535 +0000 UTC m=+153.985296130" watchObservedRunningTime="2025-12-03 21:44:17.245271238 +0000 UTC m=+153.987981833" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.245659 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.245900 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.745862513 +0000 UTC m=+154.488573108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.246672 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.251206 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.751187566 +0000 UTC m=+154.493898161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.272850 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" podStartSLOduration=129.272831136 podStartE2EDuration="2m9.272831136s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:17.270162074 +0000 UTC m=+154.012872659" watchObservedRunningTime="2025-12-03 21:44:17.272831136 +0000 UTC m=+154.015541731" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.302205 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-bldms" podStartSLOduration=129.302185984 podStartE2EDuration="2m9.302185984s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:17.300449087 +0000 UTC m=+154.043159682" watchObservedRunningTime="2025-12-03 21:44:17.302185984 +0000 UTC m=+154.044896579" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.349146 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.349389 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.849354539 +0000 UTC m=+154.592065134 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.349967 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.350324 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.850309635 +0000 UTC m=+154.593020230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.378236 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.378216164 podStartE2EDuration="2.378216164s" podCreationTimestamp="2025-12-03 21:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:17.365804451 +0000 UTC m=+154.108515046" watchObservedRunningTime="2025-12-03 21:44:17.378216164 +0000 UTC m=+154.120926759" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.438950 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-mn5lf" podStartSLOduration=128.438925462 podStartE2EDuration="2m8.438925462s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:17.436432885 +0000 UTC m=+154.179143480" watchObservedRunningTime="2025-12-03 21:44:17.438925462 +0000 UTC m=+154.181636057" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.457617 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.458497 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:17.958471686 +0000 UTC m=+154.701182271 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.511954 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:17 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:17 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:17 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.512029 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.549659 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-29fsw" podStartSLOduration=128.549638831 podStartE2EDuration="2m8.549638831s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:17.486000894 +0000 UTC m=+154.228711489" watchObservedRunningTime="2025-12-03 21:44:17.549638831 +0000 UTC m=+154.292349426" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.560574 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.560935 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:18.060921684 +0000 UTC m=+154.803632279 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.633365 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" podStartSLOduration=128.633342247 podStartE2EDuration="2m8.633342247s" podCreationTimestamp="2025-12-03 21:42:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:17.591424382 +0000 UTC m=+154.334134977" watchObservedRunningTime="2025-12-03 21:44:17.633342247 +0000 UTC m=+154.376052842" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.637006 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gtmvb"] Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.639296 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.643486 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.663196 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.663574 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 21:44:18.163553837 +0000 UTC m=+154.906264432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.690636 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gtmvb"] Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.765465 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-utilities\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.765532 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-catalog-content\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.765559 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rljmh\" (UniqueName: \"kubernetes.io/projected/d631c483-9492-4253-8029-078d1120650d-kube-api-access-rljmh\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.765586 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:17 crc kubenswrapper[4715]: E1203 21:44:17.765929 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 21:44:18.265912032 +0000 UTC m=+155.008622627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-rpjh4" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.816154 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pm9xt"] Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.817161 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.821526 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.831611 4715 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T21:44:17.121372504Z","Handler":null,"Name":""} Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.847841 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pm9xt"] Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.868745 4715 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.868783 4715 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.869721 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.870013 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-catalog-content\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.870043 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rljmh\" (UniqueName: \"kubernetes.io/projected/d631c483-9492-4253-8029-078d1120650d-kube-api-access-rljmh\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.870095 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-utilities\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.870129 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-catalog-content\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.870162 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnzth\" (UniqueName: \"kubernetes.io/projected/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-kube-api-access-wnzth\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.870197 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-utilities\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.870662 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-utilities\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.870953 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-catalog-content\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.883539 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.899378 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rljmh\" (UniqueName: \"kubernetes.io/projected/d631c483-9492-4253-8029-078d1120650d-kube-api-access-rljmh\") pod \"certified-operators-gtmvb\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.966114 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.971459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnzth\" (UniqueName: \"kubernetes.io/projected/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-kube-api-access-wnzth\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.971566 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.971607 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-utilities\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.971652 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-catalog-content\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.972107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-catalog-content\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:17 crc kubenswrapper[4715]: I1203 21:44:17.972290 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-utilities\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.000377 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnzth\" (UniqueName: \"kubernetes.io/projected/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-kube-api-access-wnzth\") pod \"community-operators-pm9xt\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.015084 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sxhqx"] Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.016695 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.033231 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sxhqx"] Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.038813 4715 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.038864 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.073209 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzp5b\" (UniqueName: \"kubernetes.io/projected/42e6daca-1b1b-4179-bf73-f5d1b6c34542-kube-api-access-lzp5b\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.073620 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-catalog-content\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.073701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-utilities\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.130151 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.186791 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-catalog-content\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.187138 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-utilities\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.187342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzp5b\" (UniqueName: \"kubernetes.io/projected/42e6daca-1b1b-4179-bf73-f5d1b6c34542-kube-api-access-lzp5b\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.189671 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-rpjh4\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.190444 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-utilities\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.197553 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-catalog-content\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.217645 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kbl2l"] Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.220745 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.252691 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzp5b\" (UniqueName: \"kubernetes.io/projected/42e6daca-1b1b-4179-bf73-f5d1b6c34542-kube-api-access-lzp5b\") pod \"certified-operators-sxhqx\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.253104 4715 generic.go:334] "Generic (PLEG): container finished" podID="2977a90b-71d1-450a-9551-0d375dc5d691" containerID="24c38ed5921f996fc2ff25840b8f3a923d76a809a3d3f41526dc3c3fe27c8fa7" exitCode=0 Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.253378 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2977a90b-71d1-450a-9551-0d375dc5d691","Type":"ContainerDied","Data":"24c38ed5921f996fc2ff25840b8f3a923d76a809a3d3f41526dc3c3fe27c8fa7"} Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.257869 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.262453 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbl2l"] Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.297821 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" event={"ID":"e39f2c0f-2634-4192-a66d-5f283f7f8e57","Type":"ContainerStarted","Data":"e0b1fb55e4cc8285fcb1d6578c73fecf7ea28e9b9a17e46f5296fb0ba968eb28"} Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.297880 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" event={"ID":"e39f2c0f-2634-4192-a66d-5f283f7f8e57","Type":"ContainerStarted","Data":"1b10788071e7a3fd6627a735eb296d8947660deeadf050c30a8e38f36c0903de"} Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.305178 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xwrk\" (UniqueName: \"kubernetes.io/projected/bffada79-8946-452f-9c4b-de82e895a026-kube-api-access-6xwrk\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.305291 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-utilities\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.307879 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-catalog-content\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.339130 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.348698 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.381512 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-xrj4r" podStartSLOduration=10.381471864 podStartE2EDuration="10.381471864s" podCreationTimestamp="2025-12-03 21:44:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:18.367361176 +0000 UTC m=+155.110071771" watchObservedRunningTime="2025-12-03 21:44:18.381471864 +0000 UTC m=+155.124182459" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.416171 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xwrk\" (UniqueName: \"kubernetes.io/projected/bffada79-8946-452f-9c4b-de82e895a026-kube-api-access-6xwrk\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.419691 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-utilities\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.419900 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-catalog-content\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.423168 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-utilities\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.426909 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-catalog-content\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.462072 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xwrk\" (UniqueName: \"kubernetes.io/projected/bffada79-8946-452f-9c4b-de82e895a026-kube-api-access-6xwrk\") pod \"community-operators-kbl2l\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.467769 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gtmvb"] Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.480436 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:18 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:18 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:18 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.480487 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.562149 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.778433 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pm9xt"] Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.924245 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rpjh4"] Dec 03 21:44:18 crc kubenswrapper[4715]: I1203 21:44:18.990554 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sxhqx"] Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.238142 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kbl2l"] Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.322568 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" event={"ID":"4fb85d9e-e6ba-49ba-aabb-89d14be69e41","Type":"ContainerStarted","Data":"2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16"} Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.322621 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" event={"ID":"4fb85d9e-e6ba-49ba-aabb-89d14be69e41","Type":"ContainerStarted","Data":"225c5788b0515c240a562b1ef7e3476c9dceb9bde2d98b0ce608870563a97e8d"} Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.323662 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.328229 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm9xt" event={"ID":"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5","Type":"ContainerStarted","Data":"ba96be179345fea66bf433b22f8162333b456e057cf18c836cc1ae377b4b9924"} Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.330742 4715 generic.go:334] "Generic (PLEG): container finished" podID="d631c483-9492-4253-8029-078d1120650d" containerID="90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3" exitCode=0 Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.330832 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtmvb" event={"ID":"d631c483-9492-4253-8029-078d1120650d","Type":"ContainerDied","Data":"90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3"} Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.330864 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtmvb" event={"ID":"d631c483-9492-4253-8029-078d1120650d","Type":"ContainerStarted","Data":"0f669bb5230324e72f8b39b1909f8466e8e402f3ed5f839afbc34464be72bb17"} Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.332842 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.333215 4715 generic.go:334] "Generic (PLEG): container finished" podID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerID="fc80d1b1568d9240b966d06044b218ddf546c6237d7dd5633e8c397ff682e259" exitCode=0 Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.333452 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxhqx" event={"ID":"42e6daca-1b1b-4179-bf73-f5d1b6c34542","Type":"ContainerDied","Data":"fc80d1b1568d9240b966d06044b218ddf546c6237d7dd5633e8c397ff682e259"} Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.333567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxhqx" event={"ID":"42e6daca-1b1b-4179-bf73-f5d1b6c34542","Type":"ContainerStarted","Data":"a9097d9463a210a6ec2fc675f70ac3f41f2091b2cf6cb7d8f84ea050c3fde01c"} Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.393597 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" podStartSLOduration=131.393560392 podStartE2EDuration="2m11.393560392s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:19.381940291 +0000 UTC m=+156.124650876" watchObservedRunningTime="2025-12-03 21:44:19.393560392 +0000 UTC m=+156.136270987" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.484246 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:19 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:19 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:19 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.484317 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.662683 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.804244 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lb8pd"] Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.806297 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.810966 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.836241 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb8pd"] Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.878058 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.960916 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2977a90b-71d1-450a-9551-0d375dc5d691-kubelet-dir\") pod \"2977a90b-71d1-450a-9551-0d375dc5d691\" (UID: \"2977a90b-71d1-450a-9551-0d375dc5d691\") " Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.961156 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2977a90b-71d1-450a-9551-0d375dc5d691-kube-api-access\") pod \"2977a90b-71d1-450a-9551-0d375dc5d691\" (UID: \"2977a90b-71d1-450a-9551-0d375dc5d691\") " Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.961315 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2977a90b-71d1-450a-9551-0d375dc5d691-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2977a90b-71d1-450a-9551-0d375dc5d691" (UID: "2977a90b-71d1-450a-9551-0d375dc5d691"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.961495 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-utilities\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.961582 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnn26\" (UniqueName: \"kubernetes.io/projected/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-kube-api-access-dnn26\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.961670 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-catalog-content\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.961761 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2977a90b-71d1-450a-9551-0d375dc5d691-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:19 crc kubenswrapper[4715]: I1203 21:44:19.970066 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2977a90b-71d1-450a-9551-0d375dc5d691-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2977a90b-71d1-450a-9551-0d375dc5d691" (UID: "2977a90b-71d1-450a-9551-0d375dc5d691"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.063678 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-utilities\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.063774 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnn26\" (UniqueName: \"kubernetes.io/projected/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-kube-api-access-dnn26\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.064369 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-catalog-content\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.064393 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-utilities\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.064435 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2977a90b-71d1-450a-9551-0d375dc5d691-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.064931 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-catalog-content\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.085688 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnn26\" (UniqueName: \"kubernetes.io/projected/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-kube-api-access-dnn26\") pod \"redhat-marketplace-lb8pd\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.166884 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.216775 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7vnkc"] Dec 03 21:44:20 crc kubenswrapper[4715]: E1203 21:44:20.217334 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2977a90b-71d1-450a-9551-0d375dc5d691" containerName="pruner" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.217358 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2977a90b-71d1-450a-9551-0d375dc5d691" containerName="pruner" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.217604 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2977a90b-71d1-450a-9551-0d375dc5d691" containerName="pruner" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.222308 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.226871 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vnkc"] Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.349892 4715 generic.go:334] "Generic (PLEG): container finished" podID="bffada79-8946-452f-9c4b-de82e895a026" containerID="7851ec9f526f295f861d38b7747e05fa94155d9d8981f338f3540b77031fa125" exitCode=0 Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.350378 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbl2l" event={"ID":"bffada79-8946-452f-9c4b-de82e895a026","Type":"ContainerDied","Data":"7851ec9f526f295f861d38b7747e05fa94155d9d8981f338f3540b77031fa125"} Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.350448 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbl2l" event={"ID":"bffada79-8946-452f-9c4b-de82e895a026","Type":"ContainerStarted","Data":"a9693e0ca9fb9c1ac8fdd7b445d6cbd087694d2c9b3dcb4bed5d7e6d32764629"} Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.355542 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"2977a90b-71d1-450a-9551-0d375dc5d691","Type":"ContainerDied","Data":"fad1aad178196b92b00f02d96c5afa8baef59b99752e299735d3080a21ee944d"} Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.355585 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fad1aad178196b92b00f02d96c5afa8baef59b99752e299735d3080a21ee944d" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.355715 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.369311 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-utilities\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.369363 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7b7h\" (UniqueName: \"kubernetes.io/projected/8322781d-b72b-4012-aa34-b4e03930cb6e-kube-api-access-s7b7h\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.369413 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-catalog-content\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.375378 4715 generic.go:334] "Generic (PLEG): container finished" podID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerID="061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5" exitCode=0 Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.376253 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm9xt" event={"ID":"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5","Type":"ContainerDied","Data":"061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5"} Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.443013 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb8pd"] Dec 03 21:44:20 crc kubenswrapper[4715]: W1203 21:44:20.456549 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podffc954fc_ce5b_48c0_89ba_ebca500a7eee.slice/crio-94d4acdfdbe5d291216b0eaf06349c3ceef4955953d33c7903e1d5bd1bda373a WatchSource:0}: Error finding container 94d4acdfdbe5d291216b0eaf06349c3ceef4955953d33c7903e1d5bd1bda373a: Status 404 returned error can't find the container with id 94d4acdfdbe5d291216b0eaf06349c3ceef4955953d33c7903e1d5bd1bda373a Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.470661 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-utilities\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.470709 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7b7h\" (UniqueName: \"kubernetes.io/projected/8322781d-b72b-4012-aa34-b4e03930cb6e-kube-api-access-s7b7h\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.470759 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-catalog-content\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.471709 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-utilities\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.472106 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-catalog-content\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.480475 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:20 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:20 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:20 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.480557 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.495381 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7b7h\" (UniqueName: \"kubernetes.io/projected/8322781d-b72b-4012-aa34-b4e03930cb6e-kube-api-access-s7b7h\") pod \"redhat-marketplace-7vnkc\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.555705 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.782426 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vnkc"] Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.806720 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-spznz"] Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.807998 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.813911 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.818754 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-spznz"] Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.879627 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-utilities\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.879784 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-catalog-content\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.879858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsrrv\" (UniqueName: \"kubernetes.io/projected/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-kube-api-access-rsrrv\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.982489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-utilities\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.982682 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-catalog-content\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.982732 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsrrv\" (UniqueName: \"kubernetes.io/projected/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-kube-api-access-rsrrv\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.985295 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-catalog-content\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:20 crc kubenswrapper[4715]: I1203 21:44:20.987056 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-utilities\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.002680 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsrrv\" (UniqueName: \"kubernetes.io/projected/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-kube-api-access-rsrrv\") pod \"redhat-operators-spznz\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.135087 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.249180 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t7slp"] Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.250357 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.253392 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t7slp"] Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.337704 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.337761 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.345199 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.373994 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-fjvxw container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.374037 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-fjvxw" podUID="e439cfd1-5b9a-48d3-9789-0365a5d14d0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.374046 4715 patch_prober.go:28] interesting pod/downloads-7954f5f757-fjvxw container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.374107 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-fjvxw" podUID="e439cfd1-5b9a-48d3-9789-0365a5d14d0f" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.388866 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-utilities\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.388938 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-catalog-content\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.389276 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwtfv\" (UniqueName: \"kubernetes.io/projected/4830567d-e7ed-4d70-96e2-d3f79846f313-kube-api-access-vwtfv\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.407735 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.407783 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.408006 4715 generic.go:334] "Generic (PLEG): container finished" podID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerID="3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99" exitCode=0 Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.408098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb8pd" event={"ID":"ffc954fc-ce5b-48c0-89ba-ebca500a7eee","Type":"ContainerDied","Data":"3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99"} Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.408134 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb8pd" event={"ID":"ffc954fc-ce5b-48c0-89ba-ebca500a7eee","Type":"ContainerStarted","Data":"94d4acdfdbe5d291216b0eaf06349c3ceef4955953d33c7903e1d5bd1bda373a"} Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.409997 4715 patch_prober.go:28] interesting pod/console-f9d7485db-sppb5 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.410040 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-sppb5" podUID="fa7301e8-1cff-40a6-a83c-d8e2db119bce" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.413036 4715 generic.go:334] "Generic (PLEG): container finished" podID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerID="9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd" exitCode=0 Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.413099 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vnkc" event={"ID":"8322781d-b72b-4012-aa34-b4e03930cb6e","Type":"ContainerDied","Data":"9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd"} Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.413141 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vnkc" event={"ID":"8322781d-b72b-4012-aa34-b4e03930cb6e","Type":"ContainerStarted","Data":"a125ead1b3e4d1f58baa042a48e17cead6130192d6ede17e60516bfc7503bea4"} Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.417120 4715 generic.go:334] "Generic (PLEG): container finished" podID="7d88e13c-7334-49a9-822b-9ac54d6bf72f" containerID="b8079d9db0601a1f5e4bc6ab874c873fc363425ea45c0846f5af709a2beb3626" exitCode=0 Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.417199 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" event={"ID":"7d88e13c-7334-49a9-822b-9ac54d6bf72f","Type":"ContainerDied","Data":"b8079d9db0601a1f5e4bc6ab874c873fc363425ea45c0846f5af709a2beb3626"} Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.427322 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6vvkt" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.484163 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.487957 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:21 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:21 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:21 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.488032 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.491033 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-utilities\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.491074 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-catalog-content\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.491670 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwtfv\" (UniqueName: \"kubernetes.io/projected/4830567d-e7ed-4d70-96e2-d3f79846f313-kube-api-access-vwtfv\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.492646 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-utilities\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.492883 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-catalog-content\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.543029 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwtfv\" (UniqueName: \"kubernetes.io/projected/4830567d-e7ed-4d70-96e2-d3f79846f313-kube-api-access-vwtfv\") pod \"redhat-operators-t7slp\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.598326 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.665040 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-spznz"] Dec 03 21:44:21 crc kubenswrapper[4715]: W1203 21:44:21.696010 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod28ef7bf3_ca1d_44d8_b647_60fd71fe9c3e.slice/crio-c423b93119b9f36ecdb4a78185f4fb822ec496651f28edcac9d3a659e68bf7c3 WatchSource:0}: Error finding container c423b93119b9f36ecdb4a78185f4fb822ec496651f28edcac9d3a659e68bf7c3: Status 404 returned error can't find the container with id c423b93119b9f36ecdb4a78185f4fb822ec496651f28edcac9d3a659e68bf7c3 Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.950610 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.952018 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.955472 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.955541 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.957573 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 21:44:21 crc kubenswrapper[4715]: I1203 21:44:21.959561 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t7slp"] Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.005211 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c9a939f-170b-4710-9621-c385a0a206ea-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9c9a939f-170b-4710-9621-c385a0a206ea\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.005298 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c9a939f-170b-4710-9621-c385a0a206ea-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9c9a939f-170b-4710-9621-c385a0a206ea\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:22 crc kubenswrapper[4715]: W1203 21:44:22.034537 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4830567d_e7ed_4d70_96e2_d3f79846f313.slice/crio-765121ad141d5357559f099e02ebfb8510dd0fb0a549b84320622ffe72d99c73 WatchSource:0}: Error finding container 765121ad141d5357559f099e02ebfb8510dd0fb0a549b84320622ffe72d99c73: Status 404 returned error can't find the container with id 765121ad141d5357559f099e02ebfb8510dd0fb0a549b84320622ffe72d99c73 Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.072763 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.073250 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.080526 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.107032 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c9a939f-170b-4710-9621-c385a0a206ea-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9c9a939f-170b-4710-9621-c385a0a206ea\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.107166 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c9a939f-170b-4710-9621-c385a0a206ea-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9c9a939f-170b-4710-9621-c385a0a206ea\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.107976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c9a939f-170b-4710-9621-c385a0a206ea-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9c9a939f-170b-4710-9621-c385a0a206ea\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.119335 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.130135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c9a939f-170b-4710-9621-c385a0a206ea-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9c9a939f-170b-4710-9621-c385a0a206ea\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.350667 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.428696 4715 generic.go:334] "Generic (PLEG): container finished" podID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerID="75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8" exitCode=0 Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.428768 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7slp" event={"ID":"4830567d-e7ed-4d70-96e2-d3f79846f313","Type":"ContainerDied","Data":"75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8"} Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.428799 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7slp" event={"ID":"4830567d-e7ed-4d70-96e2-d3f79846f313","Type":"ContainerStarted","Data":"765121ad141d5357559f099e02ebfb8510dd0fb0a549b84320622ffe72d99c73"} Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.432307 4715 generic.go:334] "Generic (PLEG): container finished" podID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerID="5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633" exitCode=0 Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.432535 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spznz" event={"ID":"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e","Type":"ContainerDied","Data":"5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633"} Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.432596 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spznz" event={"ID":"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e","Type":"ContainerStarted","Data":"c423b93119b9f36ecdb4a78185f4fb822ec496651f28edcac9d3a659e68bf7c3"} Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.436870 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-bxzdz" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.498162 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:22 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:22 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:22 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.498286 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.798712 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.825068 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d88e13c-7334-49a9-822b-9ac54d6bf72f-config-volume\") pod \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.825840 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d88e13c-7334-49a9-822b-9ac54d6bf72f-secret-volume\") pod \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.825883 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrtnn\" (UniqueName: \"kubernetes.io/projected/7d88e13c-7334-49a9-822b-9ac54d6bf72f-kube-api-access-lrtnn\") pod \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\" (UID: \"7d88e13c-7334-49a9-822b-9ac54d6bf72f\") " Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.826310 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d88e13c-7334-49a9-822b-9ac54d6bf72f-config-volume" (OuterVolumeSpecName: "config-volume") pod "7d88e13c-7334-49a9-822b-9ac54d6bf72f" (UID: "7d88e13c-7334-49a9-822b-9ac54d6bf72f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.833268 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d88e13c-7334-49a9-822b-9ac54d6bf72f-kube-api-access-lrtnn" (OuterVolumeSpecName: "kube-api-access-lrtnn") pod "7d88e13c-7334-49a9-822b-9ac54d6bf72f" (UID: "7d88e13c-7334-49a9-822b-9ac54d6bf72f"). InnerVolumeSpecName "kube-api-access-lrtnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.844184 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d88e13c-7334-49a9-822b-9ac54d6bf72f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7d88e13c-7334-49a9-822b-9ac54d6bf72f" (UID: "7d88e13c-7334-49a9-822b-9ac54d6bf72f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.889288 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.927400 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrtnn\" (UniqueName: \"kubernetes.io/projected/7d88e13c-7334-49a9-822b-9ac54d6bf72f-kube-api-access-lrtnn\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.927428 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d88e13c-7334-49a9-822b-9ac54d6bf72f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:22 crc kubenswrapper[4715]: I1203 21:44:22.927441 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d88e13c-7334-49a9-822b-9ac54d6bf72f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:23 crc kubenswrapper[4715]: I1203 21:44:23.449809 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" event={"ID":"7d88e13c-7334-49a9-822b-9ac54d6bf72f","Type":"ContainerDied","Data":"8ede262331fa124ea0cdc2b4cd88a12d50687d14d6c2892241eeb602d78016f3"} Dec 03 21:44:23 crc kubenswrapper[4715]: I1203 21:44:23.450289 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ede262331fa124ea0cdc2b4cd88a12d50687d14d6c2892241eeb602d78016f3" Dec 03 21:44:23 crc kubenswrapper[4715]: I1203 21:44:23.449912 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc" Dec 03 21:44:23 crc kubenswrapper[4715]: I1203 21:44:23.460066 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9c9a939f-170b-4710-9621-c385a0a206ea","Type":"ContainerStarted","Data":"f31dade95d9f000041a3162ef97681308a68f7ca31746005e3db5edc713902c7"} Dec 03 21:44:23 crc kubenswrapper[4715]: I1203 21:44:23.481449 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:23 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:23 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:23 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:23 crc kubenswrapper[4715]: I1203 21:44:23.481573 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:24 crc kubenswrapper[4715]: I1203 21:44:24.472813 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9c9a939f-170b-4710-9621-c385a0a206ea","Type":"ContainerStarted","Data":"14373c61e185d22ccfaee6cd3b6ce7593b33c4184620345ec31c744f0e255deb"} Dec 03 21:44:24 crc kubenswrapper[4715]: I1203 21:44:24.479977 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:24 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:24 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:24 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:24 crc kubenswrapper[4715]: I1203 21:44:24.480033 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:24 crc kubenswrapper[4715]: I1203 21:44:24.492215 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.492203775 podStartE2EDuration="3.492203775s" podCreationTimestamp="2025-12-03 21:44:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:24.491482536 +0000 UTC m=+161.234193131" watchObservedRunningTime="2025-12-03 21:44:24.492203775 +0000 UTC m=+161.234914370" Dec 03 21:44:25 crc kubenswrapper[4715]: I1203 21:44:25.479886 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:25 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:25 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:25 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:25 crc kubenswrapper[4715]: I1203 21:44:25.479965 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:25 crc kubenswrapper[4715]: I1203 21:44:25.501906 4715 generic.go:334] "Generic (PLEG): container finished" podID="9c9a939f-170b-4710-9621-c385a0a206ea" containerID="14373c61e185d22ccfaee6cd3b6ce7593b33c4184620345ec31c744f0e255deb" exitCode=0 Dec 03 21:44:25 crc kubenswrapper[4715]: I1203 21:44:25.501966 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9c9a939f-170b-4710-9621-c385a0a206ea","Type":"ContainerDied","Data":"14373c61e185d22ccfaee6cd3b6ce7593b33c4184620345ec31c744f0e255deb"} Dec 03 21:44:26 crc kubenswrapper[4715]: I1203 21:44:26.479492 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:26 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:26 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:26 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:26 crc kubenswrapper[4715]: I1203 21:44:26.479576 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:26 crc kubenswrapper[4715]: I1203 21:44:26.671577 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-xtkkr" Dec 03 21:44:26 crc kubenswrapper[4715]: I1203 21:44:26.785621 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:26 crc kubenswrapper[4715]: I1203 21:44:26.928479 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c9a939f-170b-4710-9621-c385a0a206ea-kubelet-dir\") pod \"9c9a939f-170b-4710-9621-c385a0a206ea\" (UID: \"9c9a939f-170b-4710-9621-c385a0a206ea\") " Dec 03 21:44:26 crc kubenswrapper[4715]: I1203 21:44:26.928626 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c9a939f-170b-4710-9621-c385a0a206ea-kube-api-access\") pod \"9c9a939f-170b-4710-9621-c385a0a206ea\" (UID: \"9c9a939f-170b-4710-9621-c385a0a206ea\") " Dec 03 21:44:26 crc kubenswrapper[4715]: I1203 21:44:26.929209 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9c9a939f-170b-4710-9621-c385a0a206ea-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9c9a939f-170b-4710-9621-c385a0a206ea" (UID: "9c9a939f-170b-4710-9621-c385a0a206ea"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:44:26 crc kubenswrapper[4715]: I1203 21:44:26.937766 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c9a939f-170b-4710-9621-c385a0a206ea-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9c9a939f-170b-4710-9621-c385a0a206ea" (UID: "9c9a939f-170b-4710-9621-c385a0a206ea"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:44:27 crc kubenswrapper[4715]: I1203 21:44:27.030320 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9c9a939f-170b-4710-9621-c385a0a206ea-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:27 crc kubenswrapper[4715]: I1203 21:44:27.030362 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9c9a939f-170b-4710-9621-c385a0a206ea-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:27 crc kubenswrapper[4715]: I1203 21:44:27.480963 4715 patch_prober.go:28] interesting pod/router-default-5444994796-p4drr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 21:44:27 crc kubenswrapper[4715]: [-]has-synced failed: reason withheld Dec 03 21:44:27 crc kubenswrapper[4715]: [+]process-running ok Dec 03 21:44:27 crc kubenswrapper[4715]: healthz check failed Dec 03 21:44:27 crc kubenswrapper[4715]: I1203 21:44:27.481114 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-p4drr" podUID="799b37a2-5dee-4d8d-954a-e7bd3f31bfde" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 21:44:27 crc kubenswrapper[4715]: I1203 21:44:27.538834 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9c9a939f-170b-4710-9621-c385a0a206ea","Type":"ContainerDied","Data":"f31dade95d9f000041a3162ef97681308a68f7ca31746005e3db5edc713902c7"} Dec 03 21:44:27 crc kubenswrapper[4715]: I1203 21:44:27.538884 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f31dade95d9f000041a3162ef97681308a68f7ca31746005e3db5edc713902c7" Dec 03 21:44:27 crc kubenswrapper[4715]: I1203 21:44:27.538940 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 21:44:28 crc kubenswrapper[4715]: I1203 21:44:28.482088 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:28 crc kubenswrapper[4715]: I1203 21:44:28.487618 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-p4drr" Dec 03 21:44:31 crc kubenswrapper[4715]: I1203 21:44:31.390450 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-fjvxw" Dec 03 21:44:31 crc kubenswrapper[4715]: I1203 21:44:31.410319 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:44:31 crc kubenswrapper[4715]: I1203 21:44:31.414127 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:31 crc kubenswrapper[4715]: I1203 21:44:31.419194 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-sppb5" Dec 03 21:44:31 crc kubenswrapper[4715]: I1203 21:44:31.424598 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/34ccf04b-c534-43bb-94ee-fc2e7cef68c6-metrics-certs\") pod \"network-metrics-daemon-hbsmx\" (UID: \"34ccf04b-c534-43bb-94ee-fc2e7cef68c6\") " pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:44:31 crc kubenswrapper[4715]: I1203 21:44:31.461180 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-hbsmx" Dec 03 21:44:35 crc kubenswrapper[4715]: I1203 21:44:35.069672 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-hbsmx"] Dec 03 21:44:35 crc kubenswrapper[4715]: W1203 21:44:35.081367 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34ccf04b_c534_43bb_94ee_fc2e7cef68c6.slice/crio-b12a612d371bc6c5c2a025f2c9d90f0a79f33ab44f99c1d643f58f29689b02b6 WatchSource:0}: Error finding container b12a612d371bc6c5c2a025f2c9d90f0a79f33ab44f99c1d643f58f29689b02b6: Status 404 returned error can't find the container with id b12a612d371bc6c5c2a025f2c9d90f0a79f33ab44f99c1d643f58f29689b02b6 Dec 03 21:44:35 crc kubenswrapper[4715]: I1203 21:44:35.160400 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:44:35 crc kubenswrapper[4715]: I1203 21:44:35.160493 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:44:35 crc kubenswrapper[4715]: I1203 21:44:35.604997 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" event={"ID":"34ccf04b-c534-43bb-94ee-fc2e7cef68c6","Type":"ContainerStarted","Data":"b12a612d371bc6c5c2a025f2c9d90f0a79f33ab44f99c1d643f58f29689b02b6"} Dec 03 21:44:36 crc kubenswrapper[4715]: I1203 21:44:36.613820 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" event={"ID":"34ccf04b-c534-43bb-94ee-fc2e7cef68c6","Type":"ContainerStarted","Data":"0fd11b65424d94458c4412cf606180518970855b14c0a89607d4d5f42f926cd4"} Dec 03 21:44:38 crc kubenswrapper[4715]: I1203 21:44:38.266870 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:44:46 crc kubenswrapper[4715]: I1203 21:44:46.713135 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-nmdf2_d4bdbca2-e0b6-4e6f-9749-fc76ba307440/controller-manager/0.log" Dec 03 21:44:46 crc kubenswrapper[4715]: I1203 21:44:46.713926 4715 generic.go:334] "Generic (PLEG): container finished" podID="d4bdbca2-e0b6-4e6f-9749-fc76ba307440" containerID="a4c14983b898ed988103d1d29c926152e2bae4a9c36a72197004e8d8acab466c" exitCode=137 Dec 03 21:44:46 crc kubenswrapper[4715]: I1203 21:44:46.714021 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" event={"ID":"d4bdbca2-e0b6-4e6f-9749-fc76ba307440","Type":"ContainerDied","Data":"a4c14983b898ed988103d1d29c926152e2bae4a9c36a72197004e8d8acab466c"} Dec 03 21:44:49 crc kubenswrapper[4715]: E1203 21:44:49.915893 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 21:44:49 crc kubenswrapper[4715]: E1203 21:44:49.916911 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s7b7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-7vnkc_openshift-marketplace(8322781d-b72b-4012-aa34-b4e03930cb6e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 21:44:49 crc kubenswrapper[4715]: E1203 21:44:49.918227 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-7vnkc" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" Dec 03 21:44:49 crc kubenswrapper[4715]: E1203 21:44:49.999897 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 21:44:50 crc kubenswrapper[4715]: E1203 21:44:50.000083 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rljmh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-gtmvb_openshift-marketplace(d631c483-9492-4253-8029-078d1120650d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 21:44:50 crc kubenswrapper[4715]: E1203 21:44:50.001211 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-gtmvb" podUID="d631c483-9492-4253-8029-078d1120650d" Dec 03 21:44:50 crc kubenswrapper[4715]: I1203 21:44:50.889766 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 21:44:51 crc kubenswrapper[4715]: I1203 21:44:51.851936 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-brtc2" Dec 03 21:44:53 crc kubenswrapper[4715]: I1203 21:44:53.113587 4715 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-nmdf2 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 21:44:53 crc kubenswrapper[4715]: I1203 21:44:53.113691 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" podUID="d4bdbca2-e0b6-4e6f-9749-fc76ba307440" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.731036 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 21:44:55 crc kubenswrapper[4715]: E1203 21:44:55.731880 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c9a939f-170b-4710-9621-c385a0a206ea" containerName="pruner" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.731900 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c9a939f-170b-4710-9621-c385a0a206ea" containerName="pruner" Dec 03 21:44:55 crc kubenswrapper[4715]: E1203 21:44:55.731922 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d88e13c-7334-49a9-822b-9ac54d6bf72f" containerName="collect-profiles" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.731930 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d88e13c-7334-49a9-822b-9ac54d6bf72f" containerName="collect-profiles" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.732080 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d88e13c-7334-49a9-822b-9ac54d6bf72f" containerName="collect-profiles" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.732096 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c9a939f-170b-4710-9621-c385a0a206ea" containerName="pruner" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.732724 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.735869 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.744126 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.752043 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.851949 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/523c170a-081c-4a1e-8798-dcdca863962b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"523c170a-081c-4a1e-8798-dcdca863962b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.852040 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/523c170a-081c-4a1e-8798-dcdca863962b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"523c170a-081c-4a1e-8798-dcdca863962b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.953422 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/523c170a-081c-4a1e-8798-dcdca863962b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"523c170a-081c-4a1e-8798-dcdca863962b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.953561 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/523c170a-081c-4a1e-8798-dcdca863962b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"523c170a-081c-4a1e-8798-dcdca863962b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:44:55 crc kubenswrapper[4715]: I1203 21:44:55.953722 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/523c170a-081c-4a1e-8798-dcdca863962b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"523c170a-081c-4a1e-8798-dcdca863962b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.000905 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/523c170a-081c-4a1e-8798-dcdca863962b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"523c170a-081c-4a1e-8798-dcdca863962b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.092167 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:44:56 crc kubenswrapper[4715]: E1203 21:44:56.382070 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-gtmvb" podUID="d631c483-9492-4253-8029-078d1120650d" Dec 03 21:44:56 crc kubenswrapper[4715]: E1203 21:44:56.382364 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-7vnkc" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.448119 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-nmdf2_d4bdbca2-e0b6-4e6f-9749-fc76ba307440/controller-manager/0.log" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.448223 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.489222 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zfj7z"] Dec 03 21:44:56 crc kubenswrapper[4715]: E1203 21:44:56.489622 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4bdbca2-e0b6-4e6f-9749-fc76ba307440" containerName="controller-manager" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.489648 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4bdbca2-e0b6-4e6f-9749-fc76ba307440" containerName="controller-manager" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.489814 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4bdbca2-e0b6-4e6f-9749-fc76ba307440" containerName="controller-manager" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.490409 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.493433 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zfj7z"] Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.564051 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca\") pod \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.564107 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brxm7\" (UniqueName: \"kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7\") pod \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.564163 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles\") pod \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.564208 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert\") pod \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.564232 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config\") pod \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\" (UID: \"d4bdbca2-e0b6-4e6f-9749-fc76ba307440\") " Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.565910 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca" (OuterVolumeSpecName: "client-ca") pod "d4bdbca2-e0b6-4e6f-9749-fc76ba307440" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.566049 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config" (OuterVolumeSpecName: "config") pod "d4bdbca2-e0b6-4e6f-9749-fc76ba307440" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.566133 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d4bdbca2-e0b6-4e6f-9749-fc76ba307440" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.569133 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d4bdbca2-e0b6-4e6f-9749-fc76ba307440" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.573096 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7" (OuterVolumeSpecName: "kube-api-access-brxm7") pod "d4bdbca2-e0b6-4e6f-9749-fc76ba307440" (UID: "d4bdbca2-e0b6-4e6f-9749-fc76ba307440"). InnerVolumeSpecName "kube-api-access-brxm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.665258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.665368 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-client-ca\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.665633 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-config\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.665799 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b3ac689-17bc-489d-a467-1409ab8bd635-serving-cert\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.665892 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sjw9\" (UniqueName: \"kubernetes.io/projected/8b3ac689-17bc-489d-a467-1409ab8bd635-kube-api-access-9sjw9\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.665985 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.666008 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.666026 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.666047 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.666066 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brxm7\" (UniqueName: \"kubernetes.io/projected/d4bdbca2-e0b6-4e6f-9749-fc76ba307440-kube-api-access-brxm7\") on node \"crc\" DevicePath \"\"" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.767326 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.767429 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-client-ca\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.767494 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-config\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.767575 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b3ac689-17bc-489d-a467-1409ab8bd635-serving-cert\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.767661 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sjw9\" (UniqueName: \"kubernetes.io/projected/8b3ac689-17bc-489d-a467-1409ab8bd635-kube-api-access-9sjw9\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.769587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-client-ca\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.769732 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.770084 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-config\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.774954 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b3ac689-17bc-489d-a467-1409ab8bd635-serving-cert\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.794228 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-nmdf2_d4bdbca2-e0b6-4e6f-9749-fc76ba307440/controller-manager/0.log" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.794335 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" event={"ID":"d4bdbca2-e0b6-4e6f-9749-fc76ba307440","Type":"ContainerDied","Data":"a283464962333e8b524f2825f690e035b24105123c194b06c8154747ff51af3e"} Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.794414 4715 scope.go:117] "RemoveContainer" containerID="a4c14983b898ed988103d1d29c926152e2bae4a9c36a72197004e8d8acab466c" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.794478 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-nmdf2" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.799245 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sjw9\" (UniqueName: \"kubernetes.io/projected/8b3ac689-17bc-489d-a467-1409ab8bd635-kube-api-access-9sjw9\") pod \"controller-manager-879f6c89f-zfj7z\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.805450 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.851177 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nmdf2"] Dec 03 21:44:56 crc kubenswrapper[4715]: I1203 21:44:56.856913 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-nmdf2"] Dec 03 21:44:57 crc kubenswrapper[4715]: E1203 21:44:57.331698 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 21:44:57 crc kubenswrapper[4715]: E1203 21:44:57.331923 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rsrrv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-spznz_openshift-marketplace(28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 21:44:57 crc kubenswrapper[4715]: E1203 21:44:57.333513 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-spznz" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" Dec 03 21:44:57 crc kubenswrapper[4715]: I1203 21:44:57.650238 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4bdbca2-e0b6-4e6f-9749-fc76ba307440" path="/var/lib/kubelet/pods/d4bdbca2-e0b6-4e6f-9749-fc76ba307440/volumes" Dec 03 21:44:57 crc kubenswrapper[4715]: E1203 21:44:57.807155 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-spznz" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" Dec 03 21:44:58 crc kubenswrapper[4715]: I1203 21:44:58.012349 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 21:44:58 crc kubenswrapper[4715]: I1203 21:44:58.076188 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zfj7z"] Dec 03 21:44:58 crc kubenswrapper[4715]: W1203 21:44:58.086916 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b3ac689_17bc_489d_a467_1409ab8bd635.slice/crio-ce96fa885c485d64191eae94b9546545c52708aaf5d8c53c7d6847b44dad289a WatchSource:0}: Error finding container ce96fa885c485d64191eae94b9546545c52708aaf5d8c53c7d6847b44dad289a: Status 404 returned error can't find the container with id ce96fa885c485d64191eae94b9546545c52708aaf5d8c53c7d6847b44dad289a Dec 03 21:44:58 crc kubenswrapper[4715]: I1203 21:44:58.815291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" event={"ID":"8b3ac689-17bc-489d-a467-1409ab8bd635","Type":"ContainerStarted","Data":"ce96fa885c485d64191eae94b9546545c52708aaf5d8c53c7d6847b44dad289a"} Dec 03 21:44:58 crc kubenswrapper[4715]: I1203 21:44:58.817400 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm9xt" event={"ID":"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5","Type":"ContainerStarted","Data":"5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5"} Dec 03 21:44:58 crc kubenswrapper[4715]: I1203 21:44:58.818540 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"523c170a-081c-4a1e-8798-dcdca863962b","Type":"ContainerStarted","Data":"aa617e977e84ab93d0c73645c1232f72a3776da9b29aeb8b8ec3153124ccef7e"} Dec 03 21:44:58 crc kubenswrapper[4715]: I1203 21:44:58.820088 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-hbsmx" event={"ID":"34ccf04b-c534-43bb-94ee-fc2e7cef68c6","Type":"ContainerStarted","Data":"74954ecaafc0e98439f855a49b6831b7853678b5ac2df03bd6723bdd67cef48d"} Dec 03 21:44:59 crc kubenswrapper[4715]: E1203 21:44:59.190722 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 21:44:59 crc kubenswrapper[4715]: E1203 21:44:59.190995 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vwtfv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-t7slp_openshift-marketplace(4830567d-e7ed-4d70-96e2-d3f79846f313): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 21:44:59 crc kubenswrapper[4715]: E1203 21:44:59.192291 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-t7slp" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.835658 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbl2l" event={"ID":"bffada79-8946-452f-9c4b-de82e895a026","Type":"ContainerStarted","Data":"168a593295724de450be612e4bdae623478da09770c921b6af15f301e7bcc974"} Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.843288 4715 generic.go:334] "Generic (PLEG): container finished" podID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerID="3502fcea27cf2aa004234a9bc3b5eb7ed97e1b169decc31f32b9fd97cc217bb2" exitCode=0 Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.843370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxhqx" event={"ID":"42e6daca-1b1b-4179-bf73-f5d1b6c34542","Type":"ContainerDied","Data":"3502fcea27cf2aa004234a9bc3b5eb7ed97e1b169decc31f32b9fd97cc217bb2"} Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.847533 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" event={"ID":"8b3ac689-17bc-489d-a467-1409ab8bd635","Type":"ContainerStarted","Data":"f1a91605beb814f6f6eb16be07e1ef1302620d426e96338ef5c88504db0a0bca"} Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.847845 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.850743 4715 generic.go:334] "Generic (PLEG): container finished" podID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerID="5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5" exitCode=0 Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.850828 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm9xt" event={"ID":"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5","Type":"ContainerDied","Data":"5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5"} Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.854949 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"523c170a-081c-4a1e-8798-dcdca863962b","Type":"ContainerStarted","Data":"0e93f6946f769e0994caa2ad258626dbbad91c86e79711bbb4be711d188ad956"} Dec 03 21:44:59 crc kubenswrapper[4715]: E1203 21:44:59.856643 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-t7slp" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.858949 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.944279 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" podStartSLOduration=44.944237577 podStartE2EDuration="44.944237577s" podCreationTimestamp="2025-12-03 21:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:59.934723571 +0000 UTC m=+196.677434206" watchObservedRunningTime="2025-12-03 21:44:59.944237577 +0000 UTC m=+196.686948222" Dec 03 21:44:59 crc kubenswrapper[4715]: I1203 21:44:59.987548 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-hbsmx" podStartSLOduration=171.987493957 podStartE2EDuration="2m51.987493957s" podCreationTimestamp="2025-12-03 21:42:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:44:59.961327695 +0000 UTC m=+196.704038280" watchObservedRunningTime="2025-12-03 21:44:59.987493957 +0000 UTC m=+196.730204592" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.138855 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs"] Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.139693 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.141791 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.141999 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.154238 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs"] Dec 03 21:45:00 crc kubenswrapper[4715]: E1203 21:45:00.196564 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 21:45:00 crc kubenswrapper[4715]: E1203 21:45:00.196812 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dnn26,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-lb8pd_openshift-marketplace(ffc954fc-ce5b-48c0-89ba-ebca500a7eee): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 21:45:00 crc kubenswrapper[4715]: E1203 21:45:00.198091 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-lb8pd" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.228613 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55aa040e-0a60-41b2-add8-f2900df05cdb-secret-volume\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.228681 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79l7t\" (UniqueName: \"kubernetes.io/projected/55aa040e-0a60-41b2-add8-f2900df05cdb-kube-api-access-79l7t\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.228860 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55aa040e-0a60-41b2-add8-f2900df05cdb-config-volume\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.329815 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55aa040e-0a60-41b2-add8-f2900df05cdb-secret-volume\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.329902 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79l7t\" (UniqueName: \"kubernetes.io/projected/55aa040e-0a60-41b2-add8-f2900df05cdb-kube-api-access-79l7t\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.329936 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55aa040e-0a60-41b2-add8-f2900df05cdb-config-volume\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.330976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55aa040e-0a60-41b2-add8-f2900df05cdb-config-volume\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.337289 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55aa040e-0a60-41b2-add8-f2900df05cdb-secret-volume\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.360697 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79l7t\" (UniqueName: \"kubernetes.io/projected/55aa040e-0a60-41b2-add8-f2900df05cdb-kube-api-access-79l7t\") pod \"collect-profiles-29413305-5xcgs\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.473120 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.728560 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.729892 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.735043 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs"] Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.736582 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kube-api-access\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.736644 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-var-lock\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.736940 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.738545 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 21:45:00 crc kubenswrapper[4715]: W1203 21:45:00.739634 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55aa040e_0a60_41b2_add8_f2900df05cdb.slice/crio-3b5138e83b5dbf42fe9859d486f5d5b65b5226fcfd44ef48107fbfaabdd060d4 WatchSource:0}: Error finding container 3b5138e83b5dbf42fe9859d486f5d5b65b5226fcfd44ef48107fbfaabdd060d4: Status 404 returned error can't find the container with id 3b5138e83b5dbf42fe9859d486f5d5b65b5226fcfd44ef48107fbfaabdd060d4 Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.838205 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.838296 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kube-api-access\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.838323 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-var-lock\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.838345 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kubelet-dir\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.838437 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-var-lock\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.869932 4715 generic.go:334] "Generic (PLEG): container finished" podID="523c170a-081c-4a1e-8798-dcdca863962b" containerID="0e93f6946f769e0994caa2ad258626dbbad91c86e79711bbb4be711d188ad956" exitCode=0 Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.870018 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"523c170a-081c-4a1e-8798-dcdca863962b","Type":"ContainerDied","Data":"0e93f6946f769e0994caa2ad258626dbbad91c86e79711bbb4be711d188ad956"} Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.872750 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kube-api-access\") pod \"installer-9-crc\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.875917 4715 generic.go:334] "Generic (PLEG): container finished" podID="bffada79-8946-452f-9c4b-de82e895a026" containerID="168a593295724de450be612e4bdae623478da09770c921b6af15f301e7bcc974" exitCode=0 Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.875985 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbl2l" event={"ID":"bffada79-8946-452f-9c4b-de82e895a026","Type":"ContainerDied","Data":"168a593295724de450be612e4bdae623478da09770c921b6af15f301e7bcc974"} Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.877224 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" event={"ID":"55aa040e-0a60-41b2-add8-f2900df05cdb","Type":"ContainerStarted","Data":"3b5138e83b5dbf42fe9859d486f5d5b65b5226fcfd44ef48107fbfaabdd060d4"} Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.880536 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxhqx" event={"ID":"42e6daca-1b1b-4179-bf73-f5d1b6c34542","Type":"ContainerStarted","Data":"5e9384320beb315c2ca606f459d9d46da4f9f3eec5a5110d5018214c1633b9f0"} Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.883471 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm9xt" event={"ID":"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5","Type":"ContainerStarted","Data":"399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42"} Dec 03 21:45:00 crc kubenswrapper[4715]: E1203 21:45:00.884640 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-lb8pd" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.928491 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sxhqx" podStartSLOduration=2.983043907 podStartE2EDuration="43.92846564s" podCreationTimestamp="2025-12-03 21:44:17 +0000 UTC" firstStartedPulling="2025-12-03 21:44:19.338892626 +0000 UTC m=+156.081603221" lastFinishedPulling="2025-12-03 21:45:00.284314349 +0000 UTC m=+197.027024954" observedRunningTime="2025-12-03 21:45:00.905356494 +0000 UTC m=+197.648067099" watchObservedRunningTime="2025-12-03 21:45:00.92846564 +0000 UTC m=+197.671176245" Dec 03 21:45:00 crc kubenswrapper[4715]: I1203 21:45:00.945621 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pm9xt" podStartSLOduration=3.979548911 podStartE2EDuration="43.945594023s" podCreationTimestamp="2025-12-03 21:44:17 +0000 UTC" firstStartedPulling="2025-12-03 21:44:20.378189593 +0000 UTC m=+157.120900188" lastFinishedPulling="2025-12-03 21:45:00.344234695 +0000 UTC m=+197.086945300" observedRunningTime="2025-12-03 21:45:00.944814392 +0000 UTC m=+197.687525017" watchObservedRunningTime="2025-12-03 21:45:00.945594023 +0000 UTC m=+197.688304628" Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.059292 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.294790 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.892336 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbl2l" event={"ID":"bffada79-8946-452f-9c4b-de82e895a026","Type":"ContainerStarted","Data":"8c258ac97d3bcedeb17551a6640484cae1dbfcc62f94eb205ebe8a1bccbc78cb"} Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.896172 4715 generic.go:334] "Generic (PLEG): container finished" podID="55aa040e-0a60-41b2-add8-f2900df05cdb" containerID="d31c7d87d9acbaaa4018c190ce98e56cff097021a072ff198f044636d13c0999" exitCode=0 Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.896259 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" event={"ID":"55aa040e-0a60-41b2-add8-f2900df05cdb","Type":"ContainerDied","Data":"d31c7d87d9acbaaa4018c190ce98e56cff097021a072ff198f044636d13c0999"} Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.899102 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fdbf1a76-04e0-47ed-bed1-b1b809e783df","Type":"ContainerStarted","Data":"6d9480334a169954b0878ef983265f05dc3de5f43aaa755ffcde9a4c9281ace2"} Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.899194 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fdbf1a76-04e0-47ed-bed1-b1b809e783df","Type":"ContainerStarted","Data":"a59b3a659ee43ec8a42bdbb690e688a9555216503dbd82fbe8e6231357688642"} Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.923885 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kbl2l" podStartSLOduration=2.746746825 podStartE2EDuration="43.92386434s" podCreationTimestamp="2025-12-03 21:44:18 +0000 UTC" firstStartedPulling="2025-12-03 21:44:20.352567576 +0000 UTC m=+157.095278171" lastFinishedPulling="2025-12-03 21:45:01.529685091 +0000 UTC m=+198.272395686" observedRunningTime="2025-12-03 21:45:01.920071017 +0000 UTC m=+198.662781652" watchObservedRunningTime="2025-12-03 21:45:01.92386434 +0000 UTC m=+198.666574945" Dec 03 21:45:01 crc kubenswrapper[4715]: I1203 21:45:01.968766 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.968739974 podStartE2EDuration="1.968739974s" podCreationTimestamp="2025-12-03 21:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:45:01.962948377 +0000 UTC m=+198.705658982" watchObservedRunningTime="2025-12-03 21:45:01.968739974 +0000 UTC m=+198.711450579" Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.182328 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.366221 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/523c170a-081c-4a1e-8798-dcdca863962b-kube-api-access\") pod \"523c170a-081c-4a1e-8798-dcdca863962b\" (UID: \"523c170a-081c-4a1e-8798-dcdca863962b\") " Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.366371 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/523c170a-081c-4a1e-8798-dcdca863962b-kubelet-dir\") pod \"523c170a-081c-4a1e-8798-dcdca863962b\" (UID: \"523c170a-081c-4a1e-8798-dcdca863962b\") " Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.366520 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/523c170a-081c-4a1e-8798-dcdca863962b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "523c170a-081c-4a1e-8798-dcdca863962b" (UID: "523c170a-081c-4a1e-8798-dcdca863962b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.366886 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/523c170a-081c-4a1e-8798-dcdca863962b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.376162 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/523c170a-081c-4a1e-8798-dcdca863962b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "523c170a-081c-4a1e-8798-dcdca863962b" (UID: "523c170a-081c-4a1e-8798-dcdca863962b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.467735 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/523c170a-081c-4a1e-8798-dcdca863962b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.907323 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"523c170a-081c-4a1e-8798-dcdca863962b","Type":"ContainerDied","Data":"aa617e977e84ab93d0c73645c1232f72a3776da9b29aeb8b8ec3153124ccef7e"} Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.907391 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa617e977e84ab93d0c73645c1232f72a3776da9b29aeb8b8ec3153124ccef7e" Dec 03 21:45:02 crc kubenswrapper[4715]: I1203 21:45:02.907690 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.183089 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.277847 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79l7t\" (UniqueName: \"kubernetes.io/projected/55aa040e-0a60-41b2-add8-f2900df05cdb-kube-api-access-79l7t\") pod \"55aa040e-0a60-41b2-add8-f2900df05cdb\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.277925 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55aa040e-0a60-41b2-add8-f2900df05cdb-secret-volume\") pod \"55aa040e-0a60-41b2-add8-f2900df05cdb\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.277947 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55aa040e-0a60-41b2-add8-f2900df05cdb-config-volume\") pod \"55aa040e-0a60-41b2-add8-f2900df05cdb\" (UID: \"55aa040e-0a60-41b2-add8-f2900df05cdb\") " Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.278980 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55aa040e-0a60-41b2-add8-f2900df05cdb-config-volume" (OuterVolumeSpecName: "config-volume") pod "55aa040e-0a60-41b2-add8-f2900df05cdb" (UID: "55aa040e-0a60-41b2-add8-f2900df05cdb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.284077 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55aa040e-0a60-41b2-add8-f2900df05cdb-kube-api-access-79l7t" (OuterVolumeSpecName: "kube-api-access-79l7t") pod "55aa040e-0a60-41b2-add8-f2900df05cdb" (UID: "55aa040e-0a60-41b2-add8-f2900df05cdb"). InnerVolumeSpecName "kube-api-access-79l7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.284124 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55aa040e-0a60-41b2-add8-f2900df05cdb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "55aa040e-0a60-41b2-add8-f2900df05cdb" (UID: "55aa040e-0a60-41b2-add8-f2900df05cdb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.380306 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79l7t\" (UniqueName: \"kubernetes.io/projected/55aa040e-0a60-41b2-add8-f2900df05cdb-kube-api-access-79l7t\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.380373 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/55aa040e-0a60-41b2-add8-f2900df05cdb-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.380395 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/55aa040e-0a60-41b2-add8-f2900df05cdb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.925136 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" event={"ID":"55aa040e-0a60-41b2-add8-f2900df05cdb","Type":"ContainerDied","Data":"3b5138e83b5dbf42fe9859d486f5d5b65b5226fcfd44ef48107fbfaabdd060d4"} Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.925294 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b5138e83b5dbf42fe9859d486f5d5b65b5226fcfd44ef48107fbfaabdd060d4" Dec 03 21:45:03 crc kubenswrapper[4715]: I1203 21:45:03.926629 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs" Dec 03 21:45:05 crc kubenswrapper[4715]: I1203 21:45:05.160592 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:45:05 crc kubenswrapper[4715]: I1203 21:45:05.161398 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:45:08 crc kubenswrapper[4715]: I1203 21:45:08.131935 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:45:08 crc kubenswrapper[4715]: I1203 21:45:08.132010 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:45:08 crc kubenswrapper[4715]: I1203 21:45:08.343093 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:45:08 crc kubenswrapper[4715]: I1203 21:45:08.349586 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:45:08 crc kubenswrapper[4715]: I1203 21:45:08.563869 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:45:08 crc kubenswrapper[4715]: I1203 21:45:08.563972 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:45:12 crc kubenswrapper[4715]: I1203 21:45:12.032834 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:45:12 crc kubenswrapper[4715]: I1203 21:45:12.036118 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:45:12 crc kubenswrapper[4715]: I1203 21:45:12.041070 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:45:12 crc kubenswrapper[4715]: I1203 21:45:12.100389 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:45:12 crc kubenswrapper[4715]: I1203 21:45:12.109954 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:45:12 crc kubenswrapper[4715]: I1203 21:45:12.116305 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:45:13 crc kubenswrapper[4715]: I1203 21:45:13.482951 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kbl2l"] Dec 03 21:45:13 crc kubenswrapper[4715]: I1203 21:45:13.997419 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kbl2l" podUID="bffada79-8946-452f-9c4b-de82e895a026" containerName="registry-server" containerID="cri-o://8c258ac97d3bcedeb17551a6640484cae1dbfcc62f94eb205ebe8a1bccbc78cb" gracePeriod=2 Dec 03 21:45:14 crc kubenswrapper[4715]: I1203 21:45:14.484399 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sxhqx"] Dec 03 21:45:14 crc kubenswrapper[4715]: I1203 21:45:14.484860 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sxhqx" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerName="registry-server" containerID="cri-o://5e9384320beb315c2ca606f459d9d46da4f9f3eec5a5110d5018214c1633b9f0" gracePeriod=2 Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.012417 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vnkc" event={"ID":"8322781d-b72b-4012-aa34-b4e03930cb6e","Type":"ContainerStarted","Data":"c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496"} Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.015533 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxhqx" event={"ID":"42e6daca-1b1b-4179-bf73-f5d1b6c34542","Type":"ContainerDied","Data":"5e9384320beb315c2ca606f459d9d46da4f9f3eec5a5110d5018214c1633b9f0"} Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.015490 4715 generic.go:334] "Generic (PLEG): container finished" podID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerID="5e9384320beb315c2ca606f459d9d46da4f9f3eec5a5110d5018214c1633b9f0" exitCode=0 Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.019940 4715 generic.go:334] "Generic (PLEG): container finished" podID="bffada79-8946-452f-9c4b-de82e895a026" containerID="8c258ac97d3bcedeb17551a6640484cae1dbfcc62f94eb205ebe8a1bccbc78cb" exitCode=0 Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.019991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbl2l" event={"ID":"bffada79-8946-452f-9c4b-de82e895a026","Type":"ContainerDied","Data":"8c258ac97d3bcedeb17551a6640484cae1dbfcc62f94eb205ebe8a1bccbc78cb"} Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.346335 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.401633 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.506768 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-catalog-content\") pod \"bffada79-8946-452f-9c4b-de82e895a026\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.506880 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xwrk\" (UniqueName: \"kubernetes.io/projected/bffada79-8946-452f-9c4b-de82e895a026-kube-api-access-6xwrk\") pod \"bffada79-8946-452f-9c4b-de82e895a026\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.506914 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-utilities\") pod \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.506958 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-utilities\") pod \"bffada79-8946-452f-9c4b-de82e895a026\" (UID: \"bffada79-8946-452f-9c4b-de82e895a026\") " Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.506988 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzp5b\" (UniqueName: \"kubernetes.io/projected/42e6daca-1b1b-4179-bf73-f5d1b6c34542-kube-api-access-lzp5b\") pod \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.507019 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-catalog-content\") pod \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\" (UID: \"42e6daca-1b1b-4179-bf73-f5d1b6c34542\") " Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.523280 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-utilities" (OuterVolumeSpecName: "utilities") pod "bffada79-8946-452f-9c4b-de82e895a026" (UID: "bffada79-8946-452f-9c4b-de82e895a026"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.523305 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-utilities" (OuterVolumeSpecName: "utilities") pod "42e6daca-1b1b-4179-bf73-f5d1b6c34542" (UID: "42e6daca-1b1b-4179-bf73-f5d1b6c34542"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.532765 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bffada79-8946-452f-9c4b-de82e895a026-kube-api-access-6xwrk" (OuterVolumeSpecName: "kube-api-access-6xwrk") pod "bffada79-8946-452f-9c4b-de82e895a026" (UID: "bffada79-8946-452f-9c4b-de82e895a026"). InnerVolumeSpecName "kube-api-access-6xwrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.534070 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42e6daca-1b1b-4179-bf73-f5d1b6c34542-kube-api-access-lzp5b" (OuterVolumeSpecName: "kube-api-access-lzp5b") pod "42e6daca-1b1b-4179-bf73-f5d1b6c34542" (UID: "42e6daca-1b1b-4179-bf73-f5d1b6c34542"). InnerVolumeSpecName "kube-api-access-lzp5b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.606660 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42e6daca-1b1b-4179-bf73-f5d1b6c34542" (UID: "42e6daca-1b1b-4179-bf73-f5d1b6c34542"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.608851 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xwrk\" (UniqueName: \"kubernetes.io/projected/bffada79-8946-452f-9c4b-de82e895a026-kube-api-access-6xwrk\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.609111 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.609126 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.609141 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzp5b\" (UniqueName: \"kubernetes.io/projected/42e6daca-1b1b-4179-bf73-f5d1b6c34542-kube-api-access-lzp5b\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.609309 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bffada79-8946-452f-9c4b-de82e895a026" (UID: "bffada79-8946-452f-9c4b-de82e895a026"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.709881 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42e6daca-1b1b-4179-bf73-f5d1b6c34542-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:16 crc kubenswrapper[4715]: I1203 21:45:16.709918 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bffada79-8946-452f-9c4b-de82e895a026-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.029156 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7slp" event={"ID":"4830567d-e7ed-4d70-96e2-d3f79846f313","Type":"ContainerStarted","Data":"fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066"} Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.035687 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kbl2l" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.036634 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kbl2l" event={"ID":"bffada79-8946-452f-9c4b-de82e895a026","Type":"ContainerDied","Data":"a9693e0ca9fb9c1ac8fdd7b445d6cbd087694d2c9b3dcb4bed5d7e6d32764629"} Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.036732 4715 scope.go:117] "RemoveContainer" containerID="8c258ac97d3bcedeb17551a6640484cae1dbfcc62f94eb205ebe8a1bccbc78cb" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.041309 4715 generic.go:334] "Generic (PLEG): container finished" podID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerID="c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496" exitCode=0 Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.041452 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vnkc" event={"ID":"8322781d-b72b-4012-aa34-b4e03930cb6e","Type":"ContainerDied","Data":"c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496"} Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.044239 4715 generic.go:334] "Generic (PLEG): container finished" podID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerID="5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75" exitCode=0 Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.044311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb8pd" event={"ID":"ffc954fc-ce5b-48c0-89ba-ebca500a7eee","Type":"ContainerDied","Data":"5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75"} Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.047410 4715 generic.go:334] "Generic (PLEG): container finished" podID="d631c483-9492-4253-8029-078d1120650d" containerID="0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb" exitCode=0 Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.047484 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtmvb" event={"ID":"d631c483-9492-4253-8029-078d1120650d","Type":"ContainerDied","Data":"0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb"} Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.054529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sxhqx" event={"ID":"42e6daca-1b1b-4179-bf73-f5d1b6c34542","Type":"ContainerDied","Data":"a9097d9463a210a6ec2fc675f70ac3f41f2091b2cf6cb7d8f84ea050c3fde01c"} Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.054572 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sxhqx" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.064637 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spznz" event={"ID":"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e","Type":"ContainerStarted","Data":"39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8"} Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.126344 4715 scope.go:117] "RemoveContainer" containerID="168a593295724de450be612e4bdae623478da09770c921b6af15f301e7bcc974" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.137430 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kbl2l"] Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.143038 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kbl2l"] Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.156995 4715 scope.go:117] "RemoveContainer" containerID="7851ec9f526f295f861d38b7747e05fa94155d9d8981f338f3540b77031fa125" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.158215 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sxhqx"] Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.164347 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sxhqx"] Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.171569 4715 scope.go:117] "RemoveContainer" containerID="5e9384320beb315c2ca606f459d9d46da4f9f3eec5a5110d5018214c1633b9f0" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.188405 4715 scope.go:117] "RemoveContainer" containerID="3502fcea27cf2aa004234a9bc3b5eb7ed97e1b169decc31f32b9fd97cc217bb2" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.235088 4715 scope.go:117] "RemoveContainer" containerID="fc80d1b1568d9240b966d06044b218ddf546c6237d7dd5633e8c397ff682e259" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.642069 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" path="/var/lib/kubelet/pods/42e6daca-1b1b-4179-bf73-f5d1b6c34542/volumes" Dec 03 21:45:17 crc kubenswrapper[4715]: I1203 21:45:17.643781 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bffada79-8946-452f-9c4b-de82e895a026" path="/var/lib/kubelet/pods/bffada79-8946-452f-9c4b-de82e895a026/volumes" Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.073887 4715 generic.go:334] "Generic (PLEG): container finished" podID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerID="39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8" exitCode=0 Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.075562 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spznz" event={"ID":"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e","Type":"ContainerDied","Data":"39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8"} Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.084649 4715 generic.go:334] "Generic (PLEG): container finished" podID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerID="fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066" exitCode=0 Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.084823 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7slp" event={"ID":"4830567d-e7ed-4d70-96e2-d3f79846f313","Type":"ContainerDied","Data":"fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066"} Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.084991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7slp" event={"ID":"4830567d-e7ed-4d70-96e2-d3f79846f313","Type":"ContainerStarted","Data":"f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6"} Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.093656 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vnkc" event={"ID":"8322781d-b72b-4012-aa34-b4e03930cb6e","Type":"ContainerStarted","Data":"99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc"} Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.097078 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb8pd" event={"ID":"ffc954fc-ce5b-48c0-89ba-ebca500a7eee","Type":"ContainerStarted","Data":"0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555"} Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.099829 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtmvb" event={"ID":"d631c483-9492-4253-8029-078d1120650d","Type":"ContainerStarted","Data":"fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5"} Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.118958 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t7slp" podStartSLOduration=2.076460514 podStartE2EDuration="57.118934991s" podCreationTimestamp="2025-12-03 21:44:21 +0000 UTC" firstStartedPulling="2025-12-03 21:44:22.440479031 +0000 UTC m=+159.183189626" lastFinishedPulling="2025-12-03 21:45:17.482953508 +0000 UTC m=+214.225664103" observedRunningTime="2025-12-03 21:45:18.118068078 +0000 UTC m=+214.860778673" watchObservedRunningTime="2025-12-03 21:45:18.118934991 +0000 UTC m=+214.861645576" Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.142991 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lb8pd" podStartSLOduration=3.105982277 podStartE2EDuration="59.142968961s" podCreationTimestamp="2025-12-03 21:44:19 +0000 UTC" firstStartedPulling="2025-12-03 21:44:21.40998827 +0000 UTC m=+158.152698865" lastFinishedPulling="2025-12-03 21:45:17.446974954 +0000 UTC m=+214.189685549" observedRunningTime="2025-12-03 21:45:18.138277914 +0000 UTC m=+214.880988509" watchObservedRunningTime="2025-12-03 21:45:18.142968961 +0000 UTC m=+214.885679556" Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.188782 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7vnkc" podStartSLOduration=2.170733496 podStartE2EDuration="58.18875586s" podCreationTimestamp="2025-12-03 21:44:20 +0000 UTC" firstStartedPulling="2025-12-03 21:44:21.416216006 +0000 UTC m=+158.158926601" lastFinishedPulling="2025-12-03 21:45:17.43423837 +0000 UTC m=+214.176948965" observedRunningTime="2025-12-03 21:45:18.16693317 +0000 UTC m=+214.909643765" watchObservedRunningTime="2025-12-03 21:45:18.18875586 +0000 UTC m=+214.931466455" Dec 03 21:45:18 crc kubenswrapper[4715]: I1203 21:45:18.190471 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gtmvb" podStartSLOduration=2.959971446 podStartE2EDuration="1m1.190465256s" podCreationTimestamp="2025-12-03 21:44:17 +0000 UTC" firstStartedPulling="2025-12-03 21:44:19.332569676 +0000 UTC m=+156.075280271" lastFinishedPulling="2025-12-03 21:45:17.563063496 +0000 UTC m=+214.305774081" observedRunningTime="2025-12-03 21:45:18.186438317 +0000 UTC m=+214.929148912" watchObservedRunningTime="2025-12-03 21:45:18.190465256 +0000 UTC m=+214.933175851" Dec 03 21:45:19 crc kubenswrapper[4715]: I1203 21:45:19.111924 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spznz" event={"ID":"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e","Type":"ContainerStarted","Data":"9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9"} Dec 03 21:45:20 crc kubenswrapper[4715]: I1203 21:45:20.167838 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:45:20 crc kubenswrapper[4715]: I1203 21:45:20.168237 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:45:20 crc kubenswrapper[4715]: I1203 21:45:20.220479 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:45:20 crc kubenswrapper[4715]: I1203 21:45:20.252087 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-spznz" podStartSLOduration=3.960344595 podStartE2EDuration="1m0.252060683s" podCreationTimestamp="2025-12-03 21:44:20 +0000 UTC" firstStartedPulling="2025-12-03 21:44:22.441119508 +0000 UTC m=+159.183830103" lastFinishedPulling="2025-12-03 21:45:18.732835606 +0000 UTC m=+215.475546191" observedRunningTime="2025-12-03 21:45:19.136195783 +0000 UTC m=+215.878906378" watchObservedRunningTime="2025-12-03 21:45:20.252060683 +0000 UTC m=+216.994771288" Dec 03 21:45:20 crc kubenswrapper[4715]: I1203 21:45:20.556874 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:45:20 crc kubenswrapper[4715]: I1203 21:45:20.557562 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:45:20 crc kubenswrapper[4715]: I1203 21:45:20.633017 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:45:21 crc kubenswrapper[4715]: I1203 21:45:21.135885 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:45:21 crc kubenswrapper[4715]: I1203 21:45:21.136382 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:45:21 crc kubenswrapper[4715]: I1203 21:45:21.599987 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:45:21 crc kubenswrapper[4715]: I1203 21:45:21.600058 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:45:22 crc kubenswrapper[4715]: I1203 21:45:22.208908 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-spznz" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="registry-server" probeResult="failure" output=< Dec 03 21:45:22 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 03 21:45:22 crc kubenswrapper[4715]: > Dec 03 21:45:22 crc kubenswrapper[4715]: I1203 21:45:22.654918 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t7slp" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="registry-server" probeResult="failure" output=< Dec 03 21:45:22 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 03 21:45:22 crc kubenswrapper[4715]: > Dec 03 21:45:27 crc kubenswrapper[4715]: I1203 21:45:27.966418 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:45:27 crc kubenswrapper[4715]: I1203 21:45:27.966527 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:45:28 crc kubenswrapper[4715]: I1203 21:45:28.042987 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:45:28 crc kubenswrapper[4715]: I1203 21:45:28.244430 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:45:30 crc kubenswrapper[4715]: I1203 21:45:30.232195 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:45:30 crc kubenswrapper[4715]: I1203 21:45:30.632674 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:45:31 crc kubenswrapper[4715]: I1203 21:45:31.209431 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:45:31 crc kubenswrapper[4715]: I1203 21:45:31.287100 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:45:31 crc kubenswrapper[4715]: I1203 21:45:31.677199 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:45:31 crc kubenswrapper[4715]: I1203 21:45:31.727143 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:45:32 crc kubenswrapper[4715]: I1203 21:45:32.489066 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vnkc"] Dec 03 21:45:32 crc kubenswrapper[4715]: I1203 21:45:32.489552 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7vnkc" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerName="registry-server" containerID="cri-o://99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc" gracePeriod=2 Dec 03 21:45:34 crc kubenswrapper[4715]: I1203 21:45:34.287300 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t7slp"] Dec 03 21:45:34 crc kubenswrapper[4715]: I1203 21:45:34.288770 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t7slp" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="registry-server" containerID="cri-o://f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6" gracePeriod=2 Dec 03 21:45:35 crc kubenswrapper[4715]: I1203 21:45:35.160325 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:45:35 crc kubenswrapper[4715]: I1203 21:45:35.160445 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:45:35 crc kubenswrapper[4715]: I1203 21:45:35.160581 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:45:35 crc kubenswrapper[4715]: I1203 21:45:35.161790 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 21:45:35 crc kubenswrapper[4715]: I1203 21:45:35.162030 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6" gracePeriod=600 Dec 03 21:45:36 crc kubenswrapper[4715]: E1203 21:45:36.507920 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4830567d_e7ed_4d70_96e2_d3f79846f313.slice/crio-f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6.scope\": RecentStats: unable to find data in memory cache]" Dec 03 21:45:36 crc kubenswrapper[4715]: I1203 21:45:36.807837 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:45:36 crc kubenswrapper[4715]: I1203 21:45:36.945997 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-utilities\") pod \"8322781d-b72b-4012-aa34-b4e03930cb6e\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " Dec 03 21:45:36 crc kubenswrapper[4715]: I1203 21:45:36.947177 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-utilities" (OuterVolumeSpecName: "utilities") pod "8322781d-b72b-4012-aa34-b4e03930cb6e" (UID: "8322781d-b72b-4012-aa34-b4e03930cb6e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:45:36 crc kubenswrapper[4715]: I1203 21:45:36.947454 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7b7h\" (UniqueName: \"kubernetes.io/projected/8322781d-b72b-4012-aa34-b4e03930cb6e-kube-api-access-s7b7h\") pod \"8322781d-b72b-4012-aa34-b4e03930cb6e\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " Dec 03 21:45:36 crc kubenswrapper[4715]: I1203 21:45:36.948576 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-catalog-content\") pod \"8322781d-b72b-4012-aa34-b4e03930cb6e\" (UID: \"8322781d-b72b-4012-aa34-b4e03930cb6e\") " Dec 03 21:45:36 crc kubenswrapper[4715]: I1203 21:45:36.948886 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:36 crc kubenswrapper[4715]: I1203 21:45:36.967340 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8322781d-b72b-4012-aa34-b4e03930cb6e" (UID: "8322781d-b72b-4012-aa34-b4e03930cb6e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:45:36 crc kubenswrapper[4715]: I1203 21:45:36.980562 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8322781d-b72b-4012-aa34-b4e03930cb6e-kube-api-access-s7b7h" (OuterVolumeSpecName: "kube-api-access-s7b7h") pod "8322781d-b72b-4012-aa34-b4e03930cb6e" (UID: "8322781d-b72b-4012-aa34-b4e03930cb6e"). InnerVolumeSpecName "kube-api-access-s7b7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.049734 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7b7h\" (UniqueName: \"kubernetes.io/projected/8322781d-b72b-4012-aa34-b4e03930cb6e-kube-api-access-s7b7h\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.049780 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8322781d-b72b-4012-aa34-b4e03930cb6e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.191463 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t7slp_4830567d-e7ed-4d70-96e2-d3f79846f313/registry-server/0.log" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.193055 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.249116 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t7slp_4830567d-e7ed-4d70-96e2-d3f79846f313/registry-server/0.log" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.250350 4715 generic.go:334] "Generic (PLEG): container finished" podID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerID="f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6" exitCode=137 Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.250535 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t7slp" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.250544 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7slp" event={"ID":"4830567d-e7ed-4d70-96e2-d3f79846f313","Type":"ContainerDied","Data":"f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6"} Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.250749 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t7slp" event={"ID":"4830567d-e7ed-4d70-96e2-d3f79846f313","Type":"ContainerDied","Data":"765121ad141d5357559f099e02ebfb8510dd0fb0a549b84320622ffe72d99c73"} Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.250867 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-utilities\") pod \"4830567d-e7ed-4d70-96e2-d3f79846f313\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.250957 4715 scope.go:117] "RemoveContainer" containerID="f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.251050 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwtfv\" (UniqueName: \"kubernetes.io/projected/4830567d-e7ed-4d70-96e2-d3f79846f313-kube-api-access-vwtfv\") pod \"4830567d-e7ed-4d70-96e2-d3f79846f313\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.251088 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-catalog-content\") pod \"4830567d-e7ed-4d70-96e2-d3f79846f313\" (UID: \"4830567d-e7ed-4d70-96e2-d3f79846f313\") " Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.252044 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-utilities" (OuterVolumeSpecName: "utilities") pod "4830567d-e7ed-4d70-96e2-d3f79846f313" (UID: "4830567d-e7ed-4d70-96e2-d3f79846f313"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.253793 4715 generic.go:334] "Generic (PLEG): container finished" podID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerID="99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc" exitCode=0 Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.253859 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vnkc" event={"ID":"8322781d-b72b-4012-aa34-b4e03930cb6e","Type":"ContainerDied","Data":"99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc"} Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.253894 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7vnkc" event={"ID":"8322781d-b72b-4012-aa34-b4e03930cb6e","Type":"ContainerDied","Data":"a125ead1b3e4d1f58baa042a48e17cead6130192d6ede17e60516bfc7503bea4"} Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.253972 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7vnkc" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.256425 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4830567d-e7ed-4d70-96e2-d3f79846f313-kube-api-access-vwtfv" (OuterVolumeSpecName: "kube-api-access-vwtfv") pod "4830567d-e7ed-4d70-96e2-d3f79846f313" (UID: "4830567d-e7ed-4d70-96e2-d3f79846f313"). InnerVolumeSpecName "kube-api-access-vwtfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.259729 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6" exitCode=0 Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.259842 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6"} Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.268643 4715 scope.go:117] "RemoveContainer" containerID="fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.289959 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vnkc"] Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.293305 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7vnkc"] Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.300244 4715 scope.go:117] "RemoveContainer" containerID="75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.343964 4715 scope.go:117] "RemoveContainer" containerID="f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6" Dec 03 21:45:37 crc kubenswrapper[4715]: E1203 21:45:37.344588 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6\": container with ID starting with f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6 not found: ID does not exist" containerID="f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.344627 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6"} err="failed to get container status \"f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6\": rpc error: code = NotFound desc = could not find container \"f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6\": container with ID starting with f03370ebd97dd936ea53547489fee22e006474a4e378b77c33b0f26813a6c6b6 not found: ID does not exist" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.344656 4715 scope.go:117] "RemoveContainer" containerID="fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066" Dec 03 21:45:37 crc kubenswrapper[4715]: E1203 21:45:37.345141 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066\": container with ID starting with fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066 not found: ID does not exist" containerID="fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.345167 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066"} err="failed to get container status \"fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066\": rpc error: code = NotFound desc = could not find container \"fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066\": container with ID starting with fb4b74396c9a81236124b5853b45fb4641075c95565d04f9c5179e48495f3066 not found: ID does not exist" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.345183 4715 scope.go:117] "RemoveContainer" containerID="75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8" Dec 03 21:45:37 crc kubenswrapper[4715]: E1203 21:45:37.345474 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8\": container with ID starting with 75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8 not found: ID does not exist" containerID="75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.345517 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8"} err="failed to get container status \"75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8\": rpc error: code = NotFound desc = could not find container \"75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8\": container with ID starting with 75203b6239ece9662f4a04b76f4c7ca0dd9527b81acec93d4ec448348d587db8 not found: ID does not exist" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.345537 4715 scope.go:117] "RemoveContainer" containerID="99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.352412 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwtfv\" (UniqueName: \"kubernetes.io/projected/4830567d-e7ed-4d70-96e2-d3f79846f313-kube-api-access-vwtfv\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.352447 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.363843 4715 scope.go:117] "RemoveContainer" containerID="c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.383387 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4830567d-e7ed-4d70-96e2-d3f79846f313" (UID: "4830567d-e7ed-4d70-96e2-d3f79846f313"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.404787 4715 scope.go:117] "RemoveContainer" containerID="9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.417725 4715 scope.go:117] "RemoveContainer" containerID="99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc" Dec 03 21:45:37 crc kubenswrapper[4715]: E1203 21:45:37.418112 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc\": container with ID starting with 99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc not found: ID does not exist" containerID="99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.418192 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc"} err="failed to get container status \"99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc\": rpc error: code = NotFound desc = could not find container \"99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc\": container with ID starting with 99fc3a3f8ca76de192d23b7f19f400f2444ec640f6ea58f978a4c1f50339d5fc not found: ID does not exist" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.418243 4715 scope.go:117] "RemoveContainer" containerID="c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496" Dec 03 21:45:37 crc kubenswrapper[4715]: E1203 21:45:37.418929 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496\": container with ID starting with c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496 not found: ID does not exist" containerID="c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.418992 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496"} err="failed to get container status \"c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496\": rpc error: code = NotFound desc = could not find container \"c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496\": container with ID starting with c174bb3070aaa5fb4cf7f99e302e27e1ef2e84e7cab26085ad0559329f5bb496 not found: ID does not exist" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.419031 4715 scope.go:117] "RemoveContainer" containerID="9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd" Dec 03 21:45:37 crc kubenswrapper[4715]: E1203 21:45:37.419524 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd\": container with ID starting with 9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd not found: ID does not exist" containerID="9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.419579 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd"} err="failed to get container status \"9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd\": rpc error: code = NotFound desc = could not find container \"9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd\": container with ID starting with 9cf0c493d88246987bacc0dd73b1d527bf458f1cadcb956926c00cadfa024cbd not found: ID does not exist" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.453298 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4830567d-e7ed-4d70-96e2-d3f79846f313-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.593716 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t7slp"] Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.599591 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t7slp"] Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.641520 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" path="/var/lib/kubelet/pods/4830567d-e7ed-4d70-96e2-d3f79846f313/volumes" Dec 03 21:45:37 crc kubenswrapper[4715]: I1203 21:45:37.642317 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" path="/var/lib/kubelet/pods/8322781d-b72b-4012-aa34-b4e03930cb6e/volumes" Dec 03 21:45:38 crc kubenswrapper[4715]: I1203 21:45:38.271993 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"6efdd11196c11fe6cfe26c35fe36b832d362283f353946e4db0e4144859da84c"} Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.927787 4715 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.928685 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283" gracePeriod=15 Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.928790 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace" gracePeriod=15 Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.928834 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd" gracePeriod=15 Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.928980 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab" gracePeriod=15 Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.928896 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88" gracePeriod=15 Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.933700 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.934182 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bffada79-8946-452f-9c4b-de82e895a026" containerName="extract-content" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.934343 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bffada79-8946-452f-9c4b-de82e895a026" containerName="extract-content" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.934484 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.934651 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.934773 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.934894 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.935030 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.935146 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.935278 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.935397 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.935563 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55aa040e-0a60-41b2-add8-f2900df05cdb" containerName="collect-profiles" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.935680 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="55aa040e-0a60-41b2-add8-f2900df05cdb" containerName="collect-profiles" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.935809 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="extract-content" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.935921 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="extract-content" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.936058 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerName="extract-utilities" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.936321 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerName="extract-utilities" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.936443 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.936601 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.936721 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.936839 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.936966 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.937084 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.937204 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="extract-utilities" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.937321 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="extract-utilities" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.937437 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.937588 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.937725 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.937855 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.937973 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bffada79-8946-452f-9c4b-de82e895a026" containerName="extract-utilities" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.938082 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bffada79-8946-452f-9c4b-de82e895a026" containerName="extract-utilities" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.938206 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.938327 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.938657 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bffada79-8946-452f-9c4b-de82e895a026" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.938791 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bffada79-8946-452f-9c4b-de82e895a026" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.938918 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerName="extract-content" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.939050 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerName="extract-content" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.940214 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerName="extract-utilities" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.940385 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerName="extract-utilities" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.941037 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="523c170a-081c-4a1e-8798-dcdca863962b" containerName="pruner" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.941201 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="523c170a-081c-4a1e-8798-dcdca863962b" containerName="pruner" Dec 03 21:45:39 crc kubenswrapper[4715]: E1203 21:45:39.941331 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerName="extract-content" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.941441 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerName="extract-content" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.941777 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.941983 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8322781d-b72b-4012-aa34-b4e03930cb6e" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.942102 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.942340 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="523c170a-081c-4a1e-8798-dcdca863962b" containerName="pruner" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.942646 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.942880 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.943022 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="42e6daca-1b1b-4179-bf73-f5d1b6c34542" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.943148 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="55aa040e-0a60-41b2-add8-f2900df05cdb" containerName="collect-profiles" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.943261 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4830567d-e7ed-4d70-96e2-d3f79846f313" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.943380 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.943582 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bffada79-8946-452f-9c4b-de82e895a026" containerName="registry-server" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.944045 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.947154 4715 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.950723 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:39 crc kubenswrapper[4715]: I1203 21:45:39.957142 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 03 21:45:40 crc kubenswrapper[4715]: E1203 21:45:40.031781 4715 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.092153 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.092217 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.092269 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.092308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.092337 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.092358 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.092380 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.092408 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193430 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193529 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193614 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193639 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193663 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193706 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193896 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193928 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193955 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.193982 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.194010 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.194037 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.291594 4715 generic.go:334] "Generic (PLEG): container finished" podID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" containerID="6d9480334a169954b0878ef983265f05dc3de5f43aaa755ffcde9a4c9281ace2" exitCode=0 Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.291689 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fdbf1a76-04e0-47ed-bed1-b1b809e783df","Type":"ContainerDied","Data":"6d9480334a169954b0878ef983265f05dc3de5f43aaa755ffcde9a4c9281ace2"} Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.292934 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.295766 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.297925 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.298977 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab" exitCode=0 Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.299036 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd" exitCode=0 Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.299054 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace" exitCode=0 Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.299071 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88" exitCode=2 Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.299114 4715 scope.go:117] "RemoveContainer" containerID="2df30ed357fd2ef4a39dde091558c2b62f1e8d3437acf80903ffb2ba562b83f5" Dec 03 21:45:40 crc kubenswrapper[4715]: I1203 21:45:40.333709 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:40 crc kubenswrapper[4715]: E1203 21:45:40.390983 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dd2bd83268d71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 21:45:40.390096241 +0000 UTC m=+237.132806846,LastTimestamp:2025-12-03 21:45:40.390096241 +0000 UTC m=+237.132806846,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.313695 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.317029 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"aae31165ebce13dcbd449b9fdc51f2a4c3b8f99f7629cd6daa5aeef96b041122"} Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.317101 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a41544e7699e5860b78a18f78094f0f814bcfe1636caa3320b37616b569feb7f"} Dec 03 21:45:41 crc kubenswrapper[4715]: E1203 21:45:41.318100 4715 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.318089 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.602048 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.602751 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.716842 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kubelet-dir\") pod \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.717133 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "fdbf1a76-04e0-47ed-bed1-b1b809e783df" (UID: "fdbf1a76-04e0-47ed-bed1-b1b809e783df"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.717992 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kube-api-access\") pod \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.718129 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-var-lock\") pod \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\" (UID: \"fdbf1a76-04e0-47ed-bed1-b1b809e783df\") " Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.718177 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-var-lock" (OuterVolumeSpecName: "var-lock") pod "fdbf1a76-04e0-47ed-bed1-b1b809e783df" (UID: "fdbf1a76-04e0-47ed-bed1-b1b809e783df"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.718657 4715 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.718731 4715 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.728718 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "fdbf1a76-04e0-47ed-bed1-b1b809e783df" (UID: "fdbf1a76-04e0-47ed-bed1-b1b809e783df"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:45:41 crc kubenswrapper[4715]: I1203 21:45:41.819942 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fdbf1a76-04e0-47ed-bed1-b1b809e783df-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.326939 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"fdbf1a76-04e0-47ed-bed1-b1b809e783df","Type":"ContainerDied","Data":"a59b3a659ee43ec8a42bdbb690e688a9555216503dbd82fbe8e6231357688642"} Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.327234 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a59b3a659ee43ec8a42bdbb690e688a9555216503dbd82fbe8e6231357688642" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.327036 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.341054 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.342759 4715 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283" exitCode=0 Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.350776 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.839850 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.841927 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.843580 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.844482 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.951286 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.951449 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.951841 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.951853 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.951972 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.951960 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.952696 4715 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.952731 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:42 crc kubenswrapper[4715]: I1203 21:45:42.952748 4715 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.356797 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.361029 4715 scope.go:117] "RemoveContainer" containerID="2014a2d15f49e21f5bbc8514940ff08ddaea57ae2e1317ddb2a7688bc3ec4bab" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.361355 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.388376 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.388990 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.390028 4715 scope.go:117] "RemoveContainer" containerID="ab328819ec460eb15ac171967fc473c56e338b03bbdf7b4c48323d3d14fcd8bd" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.415607 4715 scope.go:117] "RemoveContainer" containerID="718cbcd1dbc711114a0f73e14e7bc6c95fa2a74ce16d840423328ad6fca6cace" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.440222 4715 scope.go:117] "RemoveContainer" containerID="4f6143b893f2c6fdddadb2c243ff0acef60147a37b94526e9dc34d634d159e88" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.464474 4715 scope.go:117] "RemoveContainer" containerID="088dea8873e6520f144e8dd7b522d77bf68023cb9d6b7f88942b797021e68283" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.487630 4715 scope.go:117] "RemoveContainer" containerID="00488b0f43bab2e0dd2f9096e6340cc04f6d9be03bbdf171724477cae07c2e2b" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.639451 4715 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.640052 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:43 crc kubenswrapper[4715]: I1203 21:45:43.646396 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 21:45:46 crc kubenswrapper[4715]: E1203 21:45:46.646413 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:46 crc kubenswrapper[4715]: E1203 21:45:46.648403 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:46 crc kubenswrapper[4715]: E1203 21:45:46.648974 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:46 crc kubenswrapper[4715]: E1203 21:45:46.649367 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:46 crc kubenswrapper[4715]: E1203 21:45:46.649793 4715 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:46 crc kubenswrapper[4715]: I1203 21:45:46.649832 4715 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 21:45:46 crc kubenswrapper[4715]: E1203 21:45:46.650185 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Dec 03 21:45:46 crc kubenswrapper[4715]: E1203 21:45:46.851164 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Dec 03 21:45:46 crc kubenswrapper[4715]: E1203 21:45:46.903173 4715 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dd2bd83268d71 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 21:45:40.390096241 +0000 UTC m=+237.132806846,LastTimestamp:2025-12-03 21:45:40.390096241 +0000 UTC m=+237.132806846,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 21:45:47 crc kubenswrapper[4715]: E1203 21:45:47.252488 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Dec 03 21:45:48 crc kubenswrapper[4715]: E1203 21:45:48.053790 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Dec 03 21:45:49 crc kubenswrapper[4715]: E1203 21:45:49.654561 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="3.2s" Dec 03 21:45:52 crc kubenswrapper[4715]: E1203 21:45:52.856324 4715 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="6.4s" Dec 03 21:45:53 crc kubenswrapper[4715]: I1203 21:45:53.644016 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.447980 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.448071 4715 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a" exitCode=1 Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.448122 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a"} Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.449174 4715 scope.go:117] "RemoveContainer" containerID="26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.449964 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.450638 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.633498 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.635170 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.635920 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.660605 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.660642 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:45:54 crc kubenswrapper[4715]: E1203 21:45:54.661012 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:54 crc kubenswrapper[4715]: I1203 21:45:54.661553 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.460083 4715 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="93c8defc6f2c14b2a4d1f85387f739b6664febdb18511cda0bf4b06aed37202d" exitCode=0 Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.460213 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"93c8defc6f2c14b2a4d1f85387f739b6664febdb18511cda0bf4b06aed37202d"} Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.460867 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"94488412cda52bc96e110a9add49a18c818562288ac9e09098efa8b6db4d07bb"} Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.461410 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.461447 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.462304 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:55 crc kubenswrapper[4715]: E1203 21:45:55.462300 4715 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.462845 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.469606 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.469678 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b95867f927573b1a5be80c6ed142f989f77f297a9c0c72a0117425dad265770a"} Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.471014 4715 status_manager.go:851] "Failed to get status for pod" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:55 crc kubenswrapper[4715]: I1203 21:45:55.471429 4715 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.50:6443: connect: connection refused" Dec 03 21:45:56 crc kubenswrapper[4715]: I1203 21:45:56.483346 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"15140e737343ba6a91d75cfae11ecddac1a172f2b3940623ad2b92ac43a84002"} Dec 03 21:45:56 crc kubenswrapper[4715]: I1203 21:45:56.483407 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ba0397c4ab502b47f5beed2fff0ae0e00ef25b996b68e85726e887d86aa7d8f9"} Dec 03 21:45:56 crc kubenswrapper[4715]: I1203 21:45:56.483421 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7a40cbcbf57623cb8957b0950a4628fa9dff4824a249af43baa1ef063fe1de56"} Dec 03 21:45:57 crc kubenswrapper[4715]: I1203 21:45:57.505667 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"8b2e3ed15b47b017f6caef93cb142cd255103e9853c12e06393602c5e238684d"} Dec 03 21:45:57 crc kubenswrapper[4715]: I1203 21:45:57.506123 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"acfda96ac007e4efd2636e80cd1c7ad44e72f6d81107580f23939233a7f67096"} Dec 03 21:45:57 crc kubenswrapper[4715]: I1203 21:45:57.506147 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:57 crc kubenswrapper[4715]: I1203 21:45:57.506017 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:45:57 crc kubenswrapper[4715]: I1203 21:45:57.506190 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:45:59 crc kubenswrapper[4715]: I1203 21:45:59.662558 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:59 crc kubenswrapper[4715]: I1203 21:45:59.663047 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:45:59 crc kubenswrapper[4715]: I1203 21:45:59.668432 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:45:59 crc kubenswrapper[4715]: I1203 21:45:59.668775 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 21:45:59 crc kubenswrapper[4715]: I1203 21:45:59.668861 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 21:45:59 crc kubenswrapper[4715]: I1203 21:45:59.672971 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:46:02 crc kubenswrapper[4715]: I1203 21:46:02.549576 4715 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:46:03 crc kubenswrapper[4715]: I1203 21:46:03.356129 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:46:03 crc kubenswrapper[4715]: I1203 21:46:03.552910 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:46:03 crc kubenswrapper[4715]: I1203 21:46:03.552979 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:46:03 crc kubenswrapper[4715]: I1203 21:46:03.561834 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:46:03 crc kubenswrapper[4715]: I1203 21:46:03.671727 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8161a7dd-d7ee-4342-857a-9aaa12993096" Dec 03 21:46:04 crc kubenswrapper[4715]: I1203 21:46:04.560602 4715 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:46:04 crc kubenswrapper[4715]: I1203 21:46:04.560661 4715 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4b081691-dd32-48ab-a8a5-720850285986" Dec 03 21:46:04 crc kubenswrapper[4715]: I1203 21:46:04.565938 4715 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8161a7dd-d7ee-4342-857a-9aaa12993096" Dec 03 21:46:09 crc kubenswrapper[4715]: I1203 21:46:09.668901 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 21:46:09 crc kubenswrapper[4715]: I1203 21:46:09.669761 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 21:46:12 crc kubenswrapper[4715]: I1203 21:46:12.455739 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 21:46:12 crc kubenswrapper[4715]: I1203 21:46:12.473570 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 21:46:13 crc kubenswrapper[4715]: I1203 21:46:13.232565 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 21:46:13 crc kubenswrapper[4715]: I1203 21:46:13.492971 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 21:46:13 crc kubenswrapper[4715]: I1203 21:46:13.771433 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 21:46:13 crc kubenswrapper[4715]: I1203 21:46:13.932069 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 21:46:14 crc kubenswrapper[4715]: I1203 21:46:14.378376 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 21:46:14 crc kubenswrapper[4715]: I1203 21:46:14.522391 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 21:46:14 crc kubenswrapper[4715]: I1203 21:46:14.774758 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 21:46:14 crc kubenswrapper[4715]: I1203 21:46:14.779684 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 21:46:14 crc kubenswrapper[4715]: I1203 21:46:14.824221 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.061026 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.143442 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.149576 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.212648 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.296252 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.553451 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.589609 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.624437 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.674858 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.695967 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.770489 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.847219 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.877618 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.944209 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.946378 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.949739 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 21:46:15 crc kubenswrapper[4715]: I1203 21:46:15.950107 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.017122 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.018974 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.032145 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.090835 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.222167 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.412114 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.418859 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.441378 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.554095 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.614296 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.669919 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 21:46:16 crc kubenswrapper[4715]: I1203 21:46:16.760370 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.030780 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.044828 4715 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.057764 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.061248 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.099216 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.210318 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.223118 4715 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.298966 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.314729 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.361595 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.459016 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.569490 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.590000 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.593158 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.601226 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.614212 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.698957 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.749664 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.767207 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.876492 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 21:46:17 crc kubenswrapper[4715]: I1203 21:46:17.963765 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.016983 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.022781 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.052238 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.298887 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.304228 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.385561 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.474214 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.512205 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.606424 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.610026 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.654719 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.665306 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.704388 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.839337 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.884092 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 21:46:18 crc kubenswrapper[4715]: I1203 21:46:18.969291 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.063526 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.115906 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.228919 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.237985 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.275111 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.277796 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.297998 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.300956 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.346164 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.426910 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.495556 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.537393 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.543226 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.544667 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.567491 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.668894 4715 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.668985 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.669060 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.670061 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"b95867f927573b1a5be80c6ed142f989f77f297a9c0c72a0117425dad265770a"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.670303 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://b95867f927573b1a5be80c6ed142f989f77f297a9c0c72a0117425dad265770a" gracePeriod=30 Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.695091 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.701285 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.731335 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.742837 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 21:46:19 crc kubenswrapper[4715]: I1203 21:46:19.993402 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.001554 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.059710 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.061755 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.096920 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.139869 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.156666 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.160906 4715 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.175133 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.210611 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.240894 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.376050 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.491412 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.548529 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.556649 4715 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.626843 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.648857 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.794049 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 21:46:20 crc kubenswrapper[4715]: I1203 21:46:20.891684 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.017894 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.158735 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.163994 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.180491 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.183888 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.258119 4715 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.265673 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.266038 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.275395 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.297487 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.297460694 podStartE2EDuration="19.297460694s" podCreationTimestamp="2025-12-03 21:46:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:46:21.291404149 +0000 UTC m=+278.034115124" watchObservedRunningTime="2025-12-03 21:46:21.297460694 +0000 UTC m=+278.040171289" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.328079 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.370603 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.445478 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.481922 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.482902 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.550204 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.599055 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.667894 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.676825 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.759386 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.853140 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.858213 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.858932 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 21:46:21 crc kubenswrapper[4715]: I1203 21:46:21.877847 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.011152 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.083073 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.088894 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.135322 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.147946 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.207234 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.212190 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.325784 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.400907 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.417607 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.421554 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.469020 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.487384 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.504354 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.558083 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.791924 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.797448 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.923383 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 21:46:22 crc kubenswrapper[4715]: I1203 21:46:22.982109 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.047710 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.054637 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.098749 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.105814 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.126418 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.153677 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.157035 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.157930 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.267825 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.314010 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.450870 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.450870 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.587468 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.603067 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.629304 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.650121 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.721270 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.745455 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.782234 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.787199 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.815258 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.832171 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 21:46:23 crc kubenswrapper[4715]: I1203 21:46:23.982766 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.042296 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.078930 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.181740 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.255770 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.288412 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.340938 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.351331 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.393026 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.396090 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.507542 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.528966 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.565382 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.568680 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.584259 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.598584 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.680063 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.698708 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.762729 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.809174 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.826414 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.836713 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.905907 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 21:46:24 crc kubenswrapper[4715]: I1203 21:46:24.952333 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.030827 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.082388 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.198254 4715 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.219427 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.239927 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.248999 4715 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.249398 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://aae31165ebce13dcbd449b9fdc51f2a4c3b8f99f7629cd6daa5aeef96b041122" gracePeriod=5 Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.259318 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.278068 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.306017 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.449469 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.565832 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.722837 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.825583 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.905388 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 21:46:25 crc kubenswrapper[4715]: I1203 21:46:25.909453 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.110284 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.207683 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.212146 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.216043 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.263288 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.579769 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.704941 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.774051 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.803881 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.844644 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.937264 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 21:46:26 crc kubenswrapper[4715]: I1203 21:46:26.947211 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.145124 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.264347 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.310690 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.351802 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.377828 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.479632 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.537152 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.547048 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.551110 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.625722 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.711092 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.751070 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.763289 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.803184 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 21:46:27 crc kubenswrapper[4715]: I1203 21:46:27.966793 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 21:46:28 crc kubenswrapper[4715]: I1203 21:46:28.189786 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 21:46:28 crc kubenswrapper[4715]: I1203 21:46:28.552686 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 21:46:28 crc kubenswrapper[4715]: I1203 21:46:28.618388 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.776275 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.777103 4715 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="aae31165ebce13dcbd449b9fdc51f2a4c3b8f99f7629cd6daa5aeef96b041122" exitCode=137 Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.863564 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.863724 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.982576 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.982773 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.982783 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.982873 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.982896 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.982981 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.983035 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.983130 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.983309 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.983750 4715 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.983790 4715 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.983821 4715 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.983850 4715 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:46:30 crc kubenswrapper[4715]: I1203 21:46:30.999294 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:46:31 crc kubenswrapper[4715]: I1203 21:46:31.086551 4715 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:46:31 crc kubenswrapper[4715]: I1203 21:46:31.647309 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 21:46:31 crc kubenswrapper[4715]: I1203 21:46:31.790301 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 21:46:31 crc kubenswrapper[4715]: I1203 21:46:31.790433 4715 scope.go:117] "RemoveContainer" containerID="aae31165ebce13dcbd449b9fdc51f2a4c3b8f99f7629cd6daa5aeef96b041122" Dec 03 21:46:31 crc kubenswrapper[4715]: I1203 21:46:31.790594 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 21:46:39 crc kubenswrapper[4715]: I1203 21:46:39.215730 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 21:46:39 crc kubenswrapper[4715]: I1203 21:46:39.704353 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 21:46:40 crc kubenswrapper[4715]: I1203 21:46:40.479458 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 21:46:43 crc kubenswrapper[4715]: I1203 21:46:43.852549 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 21:46:45 crc kubenswrapper[4715]: I1203 21:46:45.216500 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 21:46:47 crc kubenswrapper[4715]: I1203 21:46:47.932742 4715 generic.go:334] "Generic (PLEG): container finished" podID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerID="fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab" exitCode=0 Dec 03 21:46:47 crc kubenswrapper[4715]: I1203 21:46:47.932823 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" event={"ID":"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8","Type":"ContainerDied","Data":"fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab"} Dec 03 21:46:47 crc kubenswrapper[4715]: I1203 21:46:47.933490 4715 scope.go:117] "RemoveContainer" containerID="fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab" Dec 03 21:46:48 crc kubenswrapper[4715]: I1203 21:46:48.945422 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" event={"ID":"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8","Type":"ContainerStarted","Data":"f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd"} Dec 03 21:46:48 crc kubenswrapper[4715]: I1203 21:46:48.947037 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:46:48 crc kubenswrapper[4715]: I1203 21:46:48.950315 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:46:49 crc kubenswrapper[4715]: I1203 21:46:49.959224 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 21:46:49 crc kubenswrapper[4715]: I1203 21:46:49.967796 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 21:46:49 crc kubenswrapper[4715]: I1203 21:46:49.967893 4715 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="b95867f927573b1a5be80c6ed142f989f77f297a9c0c72a0117425dad265770a" exitCode=137 Dec 03 21:46:49 crc kubenswrapper[4715]: I1203 21:46:49.968032 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"b95867f927573b1a5be80c6ed142f989f77f297a9c0c72a0117425dad265770a"} Dec 03 21:46:49 crc kubenswrapper[4715]: I1203 21:46:49.968132 4715 scope.go:117] "RemoveContainer" containerID="26424e968603478092587fc6a590394c9b159406f6753dfbd7687a663e6e078a" Dec 03 21:46:50 crc kubenswrapper[4715]: I1203 21:46:50.602132 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 21:46:50 crc kubenswrapper[4715]: I1203 21:46:50.770642 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 21:46:50 crc kubenswrapper[4715]: I1203 21:46:50.981417 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Dec 03 21:46:50 crc kubenswrapper[4715]: I1203 21:46:50.983436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e2fa4543753045f460f6b1f559e07d9c4d6ce89d4ad2f26e710acfc3278578b1"} Dec 03 21:46:53 crc kubenswrapper[4715]: I1203 21:46:53.356470 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:46:59 crc kubenswrapper[4715]: I1203 21:46:59.667973 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:46:59 crc kubenswrapper[4715]: I1203 21:46:59.674442 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:47:00 crc kubenswrapper[4715]: I1203 21:47:00.067081 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 21:47:11 crc kubenswrapper[4715]: I1203 21:47:11.887759 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 21:48:05 crc kubenswrapper[4715]: I1203 21:48:05.160639 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:48:05 crc kubenswrapper[4715]: I1203 21:48:05.161697 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:48:35 crc kubenswrapper[4715]: I1203 21:48:35.160490 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:48:35 crc kubenswrapper[4715]: I1203 21:48:35.161276 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:49:05 crc kubenswrapper[4715]: I1203 21:49:05.160221 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:49:05 crc kubenswrapper[4715]: I1203 21:49:05.161417 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:49:05 crc kubenswrapper[4715]: I1203 21:49:05.161496 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:49:05 crc kubenswrapper[4715]: I1203 21:49:05.163010 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6efdd11196c11fe6cfe26c35fe36b832d362283f353946e4db0e4144859da84c"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 21:49:05 crc kubenswrapper[4715]: I1203 21:49:05.163124 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://6efdd11196c11fe6cfe26c35fe36b832d362283f353946e4db0e4144859da84c" gracePeriod=600 Dec 03 21:49:06 crc kubenswrapper[4715]: I1203 21:49:06.177382 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="6efdd11196c11fe6cfe26c35fe36b832d362283f353946e4db0e4144859da84c" exitCode=0 Dec 03 21:49:06 crc kubenswrapper[4715]: I1203 21:49:06.177961 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"6efdd11196c11fe6cfe26c35fe36b832d362283f353946e4db0e4144859da84c"} Dec 03 21:49:06 crc kubenswrapper[4715]: I1203 21:49:06.178009 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"c756d9392222b26696ffe87c6312788d6c3e05c052a72b153df99b8342e0dfc6"} Dec 03 21:49:06 crc kubenswrapper[4715]: I1203 21:49:06.178039 4715 scope.go:117] "RemoveContainer" containerID="f8539a067324f22e8cb93a8c74be693594031222b346b8e3f7a466ffee6c1df6" Dec 03 21:51:35 crc kubenswrapper[4715]: I1203 21:51:35.160171 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:51:35 crc kubenswrapper[4715]: I1203 21:51:35.161356 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:51:47 crc kubenswrapper[4715]: I1203 21:51:47.607678 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7n294"] Dec 03 21:52:05 crc kubenswrapper[4715]: I1203 21:52:05.160727 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:52:05 crc kubenswrapper[4715]: I1203 21:52:05.161749 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:52:12 crc kubenswrapper[4715]: I1203 21:52:12.656213 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" podUID="66d4153f-10d5-4f92-8152-79509d22e17e" containerName="oauth-openshift" containerID="cri-o://bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778" gracePeriod=15 Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.105915 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.110859 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-audit-policies\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.110910 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-provider-selection\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.110948 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-service-ca\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.110985 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-ocp-branding-template\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111018 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-idp-0-file-data\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111051 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66d4153f-10d5-4f92-8152-79509d22e17e-audit-dir\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111082 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-trusted-ca-bundle\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111112 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-serving-cert\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111141 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-error\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111168 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn6mz\" (UniqueName: \"kubernetes.io/projected/66d4153f-10d5-4f92-8152-79509d22e17e-kube-api-access-rn6mz\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111199 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-cliconfig\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111221 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-login\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111260 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-session\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.111288 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-router-certs\") pod \"66d4153f-10d5-4f92-8152-79509d22e17e\" (UID: \"66d4153f-10d5-4f92-8152-79509d22e17e\") " Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.112102 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.112336 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66d4153f-10d5-4f92-8152-79509d22e17e-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.112822 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.113307 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.115218 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.119113 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.119604 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.121093 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.122416 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.125312 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d4153f-10d5-4f92-8152-79509d22e17e-kube-api-access-rn6mz" (OuterVolumeSpecName: "kube-api-access-rn6mz") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "kube-api-access-rn6mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.128197 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.128351 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.129002 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.132173 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "66d4153f-10d5-4f92-8152-79509d22e17e" (UID: "66d4153f-10d5-4f92-8152-79509d22e17e"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.176036 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr"] Dec 03 21:52:13 crc kubenswrapper[4715]: E1203 21:52:13.176388 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d4153f-10d5-4f92-8152-79509d22e17e" containerName="oauth-openshift" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.176415 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d4153f-10d5-4f92-8152-79509d22e17e" containerName="oauth-openshift" Dec 03 21:52:13 crc kubenswrapper[4715]: E1203 21:52:13.176432 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.176441 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 21:52:13 crc kubenswrapper[4715]: E1203 21:52:13.176457 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" containerName="installer" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.176465 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" containerName="installer" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.176598 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d4153f-10d5-4f92-8152-79509d22e17e" containerName="oauth-openshift" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.176612 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.176625 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdbf1a76-04e0-47ed-bed1-b1b809e783df" containerName="installer" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.177181 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.181372 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr"] Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212289 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212368 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5l2v\" (UniqueName: \"kubernetes.io/projected/19776e0c-b7e8-4868-9f9c-8228699856c2-kube-api-access-w5l2v\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212413 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212581 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212653 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-audit-policies\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212746 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-service-ca\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212778 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-session\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212855 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/19776e0c-b7e8-4868-9f9c-8228699856c2-audit-dir\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.212979 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-router-certs\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213016 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-error\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213041 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-login\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213133 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213232 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213266 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213288 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213309 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn6mz\" (UniqueName: \"kubernetes.io/projected/66d4153f-10d5-4f92-8152-79509d22e17e-kube-api-access-rn6mz\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213330 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213350 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213371 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213392 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213419 4715 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213440 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213460 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213480 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213531 4715 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/66d4153f-10d5-4f92-8152-79509d22e17e-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.213554 4715 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/66d4153f-10d5-4f92-8152-79509d22e17e-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.314749 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.314862 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-service-ca\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.314894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-session\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.314931 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/19776e0c-b7e8-4868-9f9c-8228699856c2-audit-dir\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.314963 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315007 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-router-certs\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315033 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-error\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315059 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-login\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315095 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315103 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/19776e0c-b7e8-4868-9f9c-8228699856c2-audit-dir\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315128 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315205 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5l2v\" (UniqueName: \"kubernetes.io/projected/19776e0c-b7e8-4868-9f9c-8228699856c2-kube-api-access-w5l2v\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315228 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315297 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.315323 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-audit-policies\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.316496 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-audit-policies\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.316894 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-service-ca\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.316993 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.318571 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.319127 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-session\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.319440 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.319943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.320087 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.320225 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-error\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.320610 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-user-template-login\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.320761 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-router-certs\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.322001 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/19776e0c-b7e8-4868-9f9c-8228699856c2-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.332660 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5l2v\" (UniqueName: \"kubernetes.io/projected/19776e0c-b7e8-4868-9f9c-8228699856c2-kube-api-access-w5l2v\") pod \"oauth-openshift-7db8c9bdd4-sjrpr\" (UID: \"19776e0c-b7e8-4868-9f9c-8228699856c2\") " pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.512694 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.680382 4715 generic.go:334] "Generic (PLEG): container finished" podID="66d4153f-10d5-4f92-8152-79509d22e17e" containerID="bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778" exitCode=0 Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.680450 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.680514 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" event={"ID":"66d4153f-10d5-4f92-8152-79509d22e17e","Type":"ContainerDied","Data":"bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778"} Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.681312 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7n294" event={"ID":"66d4153f-10d5-4f92-8152-79509d22e17e","Type":"ContainerDied","Data":"165b93a19d20335daa4aa3b5097e07ee706b31b008e0ddd5bd5f47e11f475b63"} Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.681407 4715 scope.go:117] "RemoveContainer" containerID="bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.711649 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7n294"] Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.719829 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7n294"] Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.734759 4715 scope.go:117] "RemoveContainer" containerID="bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778" Dec 03 21:52:13 crc kubenswrapper[4715]: E1203 21:52:13.735494 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778\": container with ID starting with bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778 not found: ID does not exist" containerID="bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.735683 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778"} err="failed to get container status \"bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778\": rpc error: code = NotFound desc = could not find container \"bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778\": container with ID starting with bc8e7af81caf5bee80fe1ea883e235fb6fdb06451c074d9ca471ad6686a74778 not found: ID does not exist" Dec 03 21:52:13 crc kubenswrapper[4715]: I1203 21:52:13.806304 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr"] Dec 03 21:52:14 crc kubenswrapper[4715]: I1203 21:52:14.689947 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" event={"ID":"19776e0c-b7e8-4868-9f9c-8228699856c2","Type":"ContainerStarted","Data":"5ca0f2781ef306bee9039f91a7848a8c521ae5a1ceea9ea5a95bb45f0256a20b"} Dec 03 21:52:14 crc kubenswrapper[4715]: I1203 21:52:14.690588 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" event={"ID":"19776e0c-b7e8-4868-9f9c-8228699856c2","Type":"ContainerStarted","Data":"6db1d5c43e3d1cc08425e3eefa333cda1185e4d8f1a6568f6fe60842e018042e"} Dec 03 21:52:14 crc kubenswrapper[4715]: I1203 21:52:14.690621 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:14 crc kubenswrapper[4715]: I1203 21:52:14.699676 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" Dec 03 21:52:14 crc kubenswrapper[4715]: I1203 21:52:14.712816 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7db8c9bdd4-sjrpr" podStartSLOduration=27.712798508 podStartE2EDuration="27.712798508s" podCreationTimestamp="2025-12-03 21:51:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:52:14.710684711 +0000 UTC m=+631.453395326" watchObservedRunningTime="2025-12-03 21:52:14.712798508 +0000 UTC m=+631.455509103" Dec 03 21:52:15 crc kubenswrapper[4715]: I1203 21:52:15.645842 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d4153f-10d5-4f92-8152-79509d22e17e" path="/var/lib/kubelet/pods/66d4153f-10d5-4f92-8152-79509d22e17e/volumes" Dec 03 21:52:35 crc kubenswrapper[4715]: I1203 21:52:35.160330 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:52:35 crc kubenswrapper[4715]: I1203 21:52:35.161349 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:52:35 crc kubenswrapper[4715]: I1203 21:52:35.161431 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:52:35 crc kubenswrapper[4715]: I1203 21:52:35.162244 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c756d9392222b26696ffe87c6312788d6c3e05c052a72b153df99b8342e0dfc6"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 21:52:35 crc kubenswrapper[4715]: I1203 21:52:35.162397 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://c756d9392222b26696ffe87c6312788d6c3e05c052a72b153df99b8342e0dfc6" gracePeriod=600 Dec 03 21:52:35 crc kubenswrapper[4715]: I1203 21:52:35.856398 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="c756d9392222b26696ffe87c6312788d6c3e05c052a72b153df99b8342e0dfc6" exitCode=0 Dec 03 21:52:35 crc kubenswrapper[4715]: I1203 21:52:35.856411 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"c756d9392222b26696ffe87c6312788d6c3e05c052a72b153df99b8342e0dfc6"} Dec 03 21:52:35 crc kubenswrapper[4715]: I1203 21:52:35.856544 4715 scope.go:117] "RemoveContainer" containerID="6efdd11196c11fe6cfe26c35fe36b832d362283f353946e4db0e4144859da84c" Dec 03 21:52:36 crc kubenswrapper[4715]: I1203 21:52:36.870291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"40f9d0ad19f2130ff948a7243a29e414da6e6f2dd3f614df95d73b2addc6413d"} Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.678288 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bv9qb"] Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.679887 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.697154 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bv9qb"] Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.813701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-registry-tls\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.813790 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.813821 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8e8520c0-53ff-4a96-b944-34eb8423b4bd-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.813854 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8e8520c0-53ff-4a96-b944-34eb8423b4bd-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.813881 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e8520c0-53ff-4a96-b944-34eb8423b4bd-trusted-ca\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.813915 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-bound-sa-token\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.813963 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xp8h\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-kube-api-access-9xp8h\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.814049 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8e8520c0-53ff-4a96-b944-34eb8423b4bd-registry-certificates\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.858402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.915171 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8e8520c0-53ff-4a96-b944-34eb8423b4bd-registry-certificates\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.915260 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-registry-tls\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.915328 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8e8520c0-53ff-4a96-b944-34eb8423b4bd-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.915368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8e8520c0-53ff-4a96-b944-34eb8423b4bd-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.915401 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e8520c0-53ff-4a96-b944-34eb8423b4bd-trusted-ca\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.915440 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-bound-sa-token\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.915478 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xp8h\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-kube-api-access-9xp8h\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.918183 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8e8520c0-53ff-4a96-b944-34eb8423b4bd-registry-certificates\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.920552 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e8520c0-53ff-4a96-b944-34eb8423b4bd-trusted-ca\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.920697 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8e8520c0-53ff-4a96-b944-34eb8423b4bd-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.947071 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8e8520c0-53ff-4a96-b944-34eb8423b4bd-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.948455 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-registry-tls\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.956105 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xp8h\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-kube-api-access-9xp8h\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:37 crc kubenswrapper[4715]: I1203 21:53:37.961052 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8e8520c0-53ff-4a96-b944-34eb8423b4bd-bound-sa-token\") pod \"image-registry-66df7c8f76-bv9qb\" (UID: \"8e8520c0-53ff-4a96-b944-34eb8423b4bd\") " pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:38 crc kubenswrapper[4715]: I1203 21:53:38.004884 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:38 crc kubenswrapper[4715]: I1203 21:53:38.221612 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bv9qb"] Dec 03 21:53:38 crc kubenswrapper[4715]: I1203 21:53:38.314529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" event={"ID":"8e8520c0-53ff-4a96-b944-34eb8423b4bd","Type":"ContainerStarted","Data":"c097929202d7ee4d8b81cb58ce2787de421dbd4cad64c1d919845ee5538b3157"} Dec 03 21:53:39 crc kubenswrapper[4715]: I1203 21:53:39.324137 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" event={"ID":"8e8520c0-53ff-4a96-b944-34eb8423b4bd","Type":"ContainerStarted","Data":"4c279807638e2e227da002f6931bd03c870f0689b9f003cb3a9cde94484fc105"} Dec 03 21:53:39 crc kubenswrapper[4715]: I1203 21:53:39.324891 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:39 crc kubenswrapper[4715]: I1203 21:53:39.352294 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" podStartSLOduration=2.352260551 podStartE2EDuration="2.352260551s" podCreationTimestamp="2025-12-03 21:53:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:53:39.351285434 +0000 UTC m=+716.093996069" watchObservedRunningTime="2025-12-03 21:53:39.352260551 +0000 UTC m=+716.094971176" Dec 03 21:53:58 crc kubenswrapper[4715]: I1203 21:53:58.015004 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bv9qb" Dec 03 21:53:58 crc kubenswrapper[4715]: I1203 21:53:58.113138 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rpjh4"] Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.185238 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" podUID="4fb85d9e-e6ba-49ba-aabb-89d14be69e41" containerName="registry" containerID="cri-o://2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16" gracePeriod=30 Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.369413 4715 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.577822 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.647874 4715 generic.go:334] "Generic (PLEG): container finished" podID="4fb85d9e-e6ba-49ba-aabb-89d14be69e41" containerID="2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16" exitCode=0 Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.647981 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.648834 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" event={"ID":"4fb85d9e-e6ba-49ba-aabb-89d14be69e41","Type":"ContainerDied","Data":"2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16"} Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.648906 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-rpjh4" event={"ID":"4fb85d9e-e6ba-49ba-aabb-89d14be69e41","Type":"ContainerDied","Data":"225c5788b0515c240a562b1ef7e3476c9dceb9bde2d98b0ce608870563a97e8d"} Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.648936 4715 scope.go:117] "RemoveContainer" containerID="2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.675803 4715 scope.go:117] "RemoveContainer" containerID="2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16" Dec 03 21:54:23 crc kubenswrapper[4715]: E1203 21:54:23.676279 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16\": container with ID starting with 2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16 not found: ID does not exist" containerID="2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.676328 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-trusted-ca\") pod \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.676345 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16"} err="failed to get container status \"2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16\": rpc error: code = NotFound desc = could not find container \"2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16\": container with ID starting with 2ac06172bc4cd9155a9e6e8e2ddf572df2f3902c4223aa5bb8e4533e7f2ecd16 not found: ID does not exist" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.676393 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-installation-pull-secrets\") pod \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.676427 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-bound-sa-token\") pod \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.676547 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsvs4\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-kube-api-access-lsvs4\") pod \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.676576 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-certificates\") pod \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.676627 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-tls\") pod \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.676924 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.677067 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-ca-trust-extracted\") pod \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\" (UID: \"4fb85d9e-e6ba-49ba-aabb-89d14be69e41\") " Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.677211 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4fb85d9e-e6ba-49ba-aabb-89d14be69e41" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.677441 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.678207 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4fb85d9e-e6ba-49ba-aabb-89d14be69e41" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.684733 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4fb85d9e-e6ba-49ba-aabb-89d14be69e41" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.684815 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4fb85d9e-e6ba-49ba-aabb-89d14be69e41" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.685369 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4fb85d9e-e6ba-49ba-aabb-89d14be69e41" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.687912 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4fb85d9e-e6ba-49ba-aabb-89d14be69e41" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.689031 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-kube-api-access-lsvs4" (OuterVolumeSpecName: "kube-api-access-lsvs4") pod "4fb85d9e-e6ba-49ba-aabb-89d14be69e41" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41"). InnerVolumeSpecName "kube-api-access-lsvs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.706858 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4fb85d9e-e6ba-49ba-aabb-89d14be69e41" (UID: "4fb85d9e-e6ba-49ba-aabb-89d14be69e41"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.779909 4715 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.780463 4715 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.780492 4715 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.780538 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsvs4\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-kube-api-access-lsvs4\") on node \"crc\" DevicePath \"\"" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.780557 4715 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.780577 4715 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4fb85d9e-e6ba-49ba-aabb-89d14be69e41-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 21:54:23 crc kubenswrapper[4715]: I1203 21:54:23.994935 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rpjh4"] Dec 03 21:54:24 crc kubenswrapper[4715]: I1203 21:54:24.004856 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-rpjh4"] Dec 03 21:54:25 crc kubenswrapper[4715]: I1203 21:54:25.648191 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fb85d9e-e6ba-49ba-aabb-89d14be69e41" path="/var/lib/kubelet/pods/4fb85d9e-e6ba-49ba-aabb-89d14be69e41/volumes" Dec 03 21:54:35 crc kubenswrapper[4715]: I1203 21:54:35.159685 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:54:35 crc kubenswrapper[4715]: I1203 21:54:35.161759 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:55:05 crc kubenswrapper[4715]: I1203 21:55:05.160882 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:55:05 crc kubenswrapper[4715]: I1203 21:55:05.161832 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:55:35 crc kubenswrapper[4715]: I1203 21:55:35.159651 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:55:35 crc kubenswrapper[4715]: I1203 21:55:35.160858 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:55:35 crc kubenswrapper[4715]: I1203 21:55:35.160968 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:55:35 crc kubenswrapper[4715]: I1203 21:55:35.162283 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"40f9d0ad19f2130ff948a7243a29e414da6e6f2dd3f614df95d73b2addc6413d"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 21:55:35 crc kubenswrapper[4715]: I1203 21:55:35.162400 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://40f9d0ad19f2130ff948a7243a29e414da6e6f2dd3f614df95d73b2addc6413d" gracePeriod=600 Dec 03 21:55:36 crc kubenswrapper[4715]: I1203 21:55:36.071988 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="40f9d0ad19f2130ff948a7243a29e414da6e6f2dd3f614df95d73b2addc6413d" exitCode=0 Dec 03 21:55:36 crc kubenswrapper[4715]: I1203 21:55:36.072105 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"40f9d0ad19f2130ff948a7243a29e414da6e6f2dd3f614df95d73b2addc6413d"} Dec 03 21:55:36 crc kubenswrapper[4715]: I1203 21:55:36.072635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"e257c66c2cc6aa263755e73d7447050e7965b799fc91ed32403621c7f701856d"} Dec 03 21:55:36 crc kubenswrapper[4715]: I1203 21:55:36.072689 4715 scope.go:117] "RemoveContainer" containerID="c756d9392222b26696ffe87c6312788d6c3e05c052a72b153df99b8342e0dfc6" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.741440 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gtmvb"] Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.743632 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gtmvb" podUID="d631c483-9492-4253-8029-078d1120650d" containerName="registry-server" containerID="cri-o://fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5" gracePeriod=30 Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.758820 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pm9xt"] Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.759191 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pm9xt" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerName="registry-server" containerID="cri-o://399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42" gracePeriod=30 Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.779289 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-28k2p"] Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.786898 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb8pd"] Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.787215 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lb8pd" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerName="registry-server" containerID="cri-o://0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555" gracePeriod=30 Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.798349 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-spznz"] Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.798721 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-spznz" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="registry-server" containerID="cri-o://9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9" gracePeriod=30 Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.815932 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ss8p4"] Dec 03 21:56:03 crc kubenswrapper[4715]: E1203 21:56:03.816231 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fb85d9e-e6ba-49ba-aabb-89d14be69e41" containerName="registry" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.816246 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fb85d9e-e6ba-49ba-aabb-89d14be69e41" containerName="registry" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.816387 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fb85d9e-e6ba-49ba-aabb-89d14be69e41" containerName="registry" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.816915 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.834922 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ss8p4"] Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.850996 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5mqb\" (UniqueName: \"kubernetes.io/projected/62dffbaa-068c-484d-822d-51bb3bfd3105-kube-api-access-n5mqb\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.851650 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62dffbaa-068c-484d-822d-51bb3bfd3105-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.851729 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/62dffbaa-068c-484d-822d-51bb3bfd3105-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.954037 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62dffbaa-068c-484d-822d-51bb3bfd3105-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.954107 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/62dffbaa-068c-484d-822d-51bb3bfd3105-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.954142 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5mqb\" (UniqueName: \"kubernetes.io/projected/62dffbaa-068c-484d-822d-51bb3bfd3105-kube-api-access-n5mqb\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.955549 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/62dffbaa-068c-484d-822d-51bb3bfd3105-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.962896 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vvxzs"] Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.964231 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.968396 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/62dffbaa-068c-484d-822d-51bb3bfd3105-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.975093 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vvxzs"] Dec 03 21:56:03 crc kubenswrapper[4715]: I1203 21:56:03.979064 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5mqb\" (UniqueName: \"kubernetes.io/projected/62dffbaa-068c-484d-822d-51bb3bfd3105-kube-api-access-n5mqb\") pod \"marketplace-operator-79b997595-ss8p4\" (UID: \"62dffbaa-068c-484d-822d-51bb3bfd3105\") " pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.055274 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-catalog-content\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.055328 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wm65\" (UniqueName: \"kubernetes.io/projected/f7912273-d7d2-4887-8525-e0c9e20d569a-kube-api-access-5wm65\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.055650 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-utilities\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.138141 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.157808 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-catalog-content\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.157877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wm65\" (UniqueName: \"kubernetes.io/projected/f7912273-d7d2-4887-8525-e0c9e20d569a-kube-api-access-5wm65\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.158136 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-utilities\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.162899 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-catalog-content\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.164927 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-utilities\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.176030 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l88hv"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.189526 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.197515 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wm65\" (UniqueName: \"kubernetes.io/projected/f7912273-d7d2-4887-8525-e0c9e20d569a-kube-api-access-5wm65\") pod \"community-operators-vvxzs\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.203322 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l88hv"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.228389 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.237239 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.249321 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.251788 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.256205 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262100 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-catalog-content\") pod \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262338 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rljmh\" (UniqueName: \"kubernetes.io/projected/d631c483-9492-4253-8029-078d1120650d-kube-api-access-rljmh\") pod \"d631c483-9492-4253-8029-078d1120650d\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262377 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-catalog-content\") pod \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262414 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-utilities\") pod \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262447 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnzth\" (UniqueName: \"kubernetes.io/projected/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-kube-api-access-wnzth\") pod \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\" (UID: \"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262483 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-utilities\") pod \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262543 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-utilities\") pod \"d631c483-9492-4253-8029-078d1120650d\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262586 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-catalog-content\") pod \"d631c483-9492-4253-8029-078d1120650d\" (UID: \"d631c483-9492-4253-8029-078d1120650d\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262631 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsrrv\" (UniqueName: \"kubernetes.io/projected/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-kube-api-access-rsrrv\") pod \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262665 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnn26\" (UniqueName: \"kubernetes.io/projected/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-kube-api-access-dnn26\") pod \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262692 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-catalog-content\") pod \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\" (UID: \"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.262718 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-utilities\") pod \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\" (UID: \"ffc954fc-ce5b-48c0-89ba-ebca500a7eee\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.263049 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-catalog-content\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.263096 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-utilities\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.263284 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5cj5\" (UniqueName: \"kubernetes.io/projected/3eda92a8-2005-4801-befd-d1db15f1b2fa-kube-api-access-n5cj5\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.267495 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-utilities" (OuterVolumeSpecName: "utilities") pod "ffc954fc-ce5b-48c0-89ba-ebca500a7eee" (UID: "ffc954fc-ce5b-48c0-89ba-ebca500a7eee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.271758 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-kube-api-access-dnn26" (OuterVolumeSpecName: "kube-api-access-dnn26") pod "ffc954fc-ce5b-48c0-89ba-ebca500a7eee" (UID: "ffc954fc-ce5b-48c0-89ba-ebca500a7eee"). InnerVolumeSpecName "kube-api-access-dnn26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.272070 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d631c483-9492-4253-8029-078d1120650d-kube-api-access-rljmh" (OuterVolumeSpecName: "kube-api-access-rljmh") pod "d631c483-9492-4253-8029-078d1120650d" (UID: "d631c483-9492-4253-8029-078d1120650d"). InnerVolumeSpecName "kube-api-access-rljmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.274532 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-utilities" (OuterVolumeSpecName: "utilities") pod "d631c483-9492-4253-8029-078d1120650d" (UID: "d631c483-9492-4253-8029-078d1120650d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.274621 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-utilities" (OuterVolumeSpecName: "utilities") pod "28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" (UID: "28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.275288 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-utilities" (OuterVolumeSpecName: "utilities") pod "2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" (UID: "2303c745-dd7b-4ccc-b0f8-7cb25c6efea5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.277703 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-kube-api-access-rsrrv" (OuterVolumeSpecName: "kube-api-access-rsrrv") pod "28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" (UID: "28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e"). InnerVolumeSpecName "kube-api-access-rsrrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.290787 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-kube-api-access-wnzth" (OuterVolumeSpecName: "kube-api-access-wnzth") pod "2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" (UID: "2303c745-dd7b-4ccc-b0f8-7cb25c6efea5"). InnerVolumeSpecName "kube-api-access-wnzth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.312138 4715 generic.go:334] "Generic (PLEG): container finished" podID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerID="9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9" exitCode=0 Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.312254 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spznz" event={"ID":"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e","Type":"ContainerDied","Data":"9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9"} Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.312302 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spznz" event={"ID":"28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e","Type":"ContainerDied","Data":"c423b93119b9f36ecdb4a78185f4fb822ec496651f28edcac9d3a659e68bf7c3"} Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.312328 4715 scope.go:117] "RemoveContainer" containerID="9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.312493 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spznz" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.318772 4715 generic.go:334] "Generic (PLEG): container finished" podID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerID="399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42" exitCode=0 Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.318855 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm9xt" event={"ID":"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5","Type":"ContainerDied","Data":"399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42"} Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.318881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pm9xt" event={"ID":"2303c745-dd7b-4ccc-b0f8-7cb25c6efea5","Type":"ContainerDied","Data":"ba96be179345fea66bf433b22f8162333b456e057cf18c836cc1ae377b4b9924"} Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.318909 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pm9xt" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.332842 4715 generic.go:334] "Generic (PLEG): container finished" podID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerID="0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555" exitCode=0 Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.332925 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb8pd" event={"ID":"ffc954fc-ce5b-48c0-89ba-ebca500a7eee","Type":"ContainerDied","Data":"0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555"} Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.332961 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lb8pd" event={"ID":"ffc954fc-ce5b-48c0-89ba-ebca500a7eee","Type":"ContainerDied","Data":"94d4acdfdbe5d291216b0eaf06349c3ceef4955953d33c7903e1d5bd1bda373a"} Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.333063 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lb8pd" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.335920 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ffc954fc-ce5b-48c0-89ba-ebca500a7eee" (UID: "ffc954fc-ce5b-48c0-89ba-ebca500a7eee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.351680 4715 scope.go:117] "RemoveContainer" containerID="39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.355891 4715 generic.go:334] "Generic (PLEG): container finished" podID="d631c483-9492-4253-8029-078d1120650d" containerID="fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5" exitCode=0 Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.356205 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" containerID="cri-o://f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd" gracePeriod=30 Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.356544 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gtmvb" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.356631 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtmvb" event={"ID":"d631c483-9492-4253-8029-078d1120650d","Type":"ContainerDied","Data":"fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5"} Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.356678 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gtmvb" event={"ID":"d631c483-9492-4253-8029-078d1120650d","Type":"ContainerDied","Data":"0f669bb5230324e72f8b39b1909f8466e8e402f3ed5f839afbc34464be72bb17"} Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366181 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5cj5\" (UniqueName: \"kubernetes.io/projected/3eda92a8-2005-4801-befd-d1db15f1b2fa-kube-api-access-n5cj5\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366242 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-catalog-content\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-utilities\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366300 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366311 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rljmh\" (UniqueName: \"kubernetes.io/projected/d631c483-9492-4253-8029-078d1120650d-kube-api-access-rljmh\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366322 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366330 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnzth\" (UniqueName: \"kubernetes.io/projected/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-kube-api-access-wnzth\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366339 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366347 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366355 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsrrv\" (UniqueName: \"kubernetes.io/projected/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-kube-api-access-rsrrv\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366363 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnn26\" (UniqueName: \"kubernetes.io/projected/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-kube-api-access-dnn26\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366372 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ffc954fc-ce5b-48c0-89ba-ebca500a7eee-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.366783 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-utilities\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.367003 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-catalog-content\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.386433 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d631c483-9492-4253-8029-078d1120650d" (UID: "d631c483-9492-4253-8029-078d1120650d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.390348 4715 scope.go:117] "RemoveContainer" containerID="5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.391374 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5cj5\" (UniqueName: \"kubernetes.io/projected/3eda92a8-2005-4801-befd-d1db15f1b2fa-kube-api-access-n5cj5\") pod \"certified-operators-l88hv\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.406137 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" (UID: "2303c745-dd7b-4ccc-b0f8-7cb25c6efea5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.429259 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" (UID: "28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.450457 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-ss8p4"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.453581 4715 scope.go:117] "RemoveContainer" containerID="9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.453947 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9\": container with ID starting with 9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9 not found: ID does not exist" containerID="9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.453981 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9"} err="failed to get container status \"9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9\": rpc error: code = NotFound desc = could not find container \"9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9\": container with ID starting with 9148710ce8bd95a06cd2e69cc73bc9614e4d988a789cb011ad240e4b361607e9 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.454005 4715 scope.go:117] "RemoveContainer" containerID="39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.454216 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8\": container with ID starting with 39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8 not found: ID does not exist" containerID="39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.454241 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8"} err="failed to get container status \"39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8\": rpc error: code = NotFound desc = could not find container \"39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8\": container with ID starting with 39b6100cc862af5adfa9c8d57589f4845309e2319f5763d51901ad6bd810a1e8 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.454254 4715 scope.go:117] "RemoveContainer" containerID="5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.454424 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633\": container with ID starting with 5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633 not found: ID does not exist" containerID="5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.454457 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633"} err="failed to get container status \"5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633\": rpc error: code = NotFound desc = could not find container \"5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633\": container with ID starting with 5c7c561cf975a5677130c6a52c4d79be26d705703466f7be40ca0f0140d7c633 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.454478 4715 scope.go:117] "RemoveContainer" containerID="399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.467258 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.467335 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d631c483-9492-4253-8029-078d1120650d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.467358 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.482818 4715 scope.go:117] "RemoveContainer" containerID="5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.508416 4715 scope.go:117] "RemoveContainer" containerID="061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.532455 4715 scope.go:117] "RemoveContainer" containerID="399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.533202 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42\": container with ID starting with 399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42 not found: ID does not exist" containerID="399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.533248 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42"} err="failed to get container status \"399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42\": rpc error: code = NotFound desc = could not find container \"399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42\": container with ID starting with 399eb4016e6e9564dc8da35da2f614820dcf09e2aed9b8baa11b35cdc7c02b42 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.533287 4715 scope.go:117] "RemoveContainer" containerID="5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.533919 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5\": container with ID starting with 5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5 not found: ID does not exist" containerID="5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.533942 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5"} err="failed to get container status \"5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5\": rpc error: code = NotFound desc = could not find container \"5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5\": container with ID starting with 5b1b64e547c72b7d59156f5bddf8bbe93a219c053aba20c885f3957e9d8fb5e5 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.533954 4715 scope.go:117] "RemoveContainer" containerID="061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.534211 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5\": container with ID starting with 061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5 not found: ID does not exist" containerID="061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.534230 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5"} err="failed to get container status \"061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5\": rpc error: code = NotFound desc = could not find container \"061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5\": container with ID starting with 061105e5acc214c807a8f4adadb0f251e0c12bead953417cff5b63c9957973e5 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.534250 4715 scope.go:117] "RemoveContainer" containerID="0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.544688 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.549741 4715 scope.go:117] "RemoveContainer" containerID="5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.576431 4715 scope.go:117] "RemoveContainer" containerID="3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.600612 4715 scope.go:117] "RemoveContainer" containerID="0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.601979 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555\": container with ID starting with 0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555 not found: ID does not exist" containerID="0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.602030 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555"} err="failed to get container status \"0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555\": rpc error: code = NotFound desc = could not find container \"0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555\": container with ID starting with 0f12d26e09e9d22626c2db52a5345858c71b4b606953b1b60b44af87a8520555 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.602070 4715 scope.go:117] "RemoveContainer" containerID="5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.602461 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75\": container with ID starting with 5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75 not found: ID does not exist" containerID="5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.602484 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75"} err="failed to get container status \"5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75\": rpc error: code = NotFound desc = could not find container \"5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75\": container with ID starting with 5b6bc32d820e0e265527fba5018f4f7790f7223072e027402a1fb73715862a75 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.602497 4715 scope.go:117] "RemoveContainer" containerID="3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.603128 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99\": container with ID starting with 3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99 not found: ID does not exist" containerID="3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.603154 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99"} err="failed to get container status \"3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99\": rpc error: code = NotFound desc = could not find container \"3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99\": container with ID starting with 3147fc63aaec8d477fb56d5f5d0516bd46e266cd57ccee9ac92f9b8eefb79a99 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.603168 4715 scope.go:117] "RemoveContainer" containerID="fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.639330 4715 scope.go:117] "RemoveContainer" containerID="0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.667648 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pm9xt"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.680739 4715 scope.go:117] "RemoveContainer" containerID="90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.682088 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pm9xt"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.689985 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb8pd"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.697461 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lb8pd"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.706376 4715 scope.go:117] "RemoveContainer" containerID="fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.706383 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-spznz"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.711915 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-spznz"] Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.715859 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5\": container with ID starting with fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5 not found: ID does not exist" containerID="fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.715898 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5"} err="failed to get container status \"fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5\": rpc error: code = NotFound desc = could not find container \"fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5\": container with ID starting with fb2b60c1c2de8b69fdecf4bfcac1134112a0fb4e3935fca05e435685443039d5 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.715930 4715 scope.go:117] "RemoveContainer" containerID="0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.717593 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb\": container with ID starting with 0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb not found: ID does not exist" containerID="0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.717773 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb"} err="failed to get container status \"0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb\": rpc error: code = NotFound desc = could not find container \"0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb\": container with ID starting with 0e5fefdedfeb231745a829dc458670b8455d4350b96ff9109a9a91538635e7fb not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.717930 4715 scope.go:117] "RemoveContainer" containerID="90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3" Dec 03 21:56:04 crc kubenswrapper[4715]: E1203 21:56:04.719688 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3\": container with ID starting with 90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3 not found: ID does not exist" containerID="90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.719724 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3"} err="failed to get container status \"90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3\": rpc error: code = NotFound desc = could not find container \"90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3\": container with ID starting with 90dd4ae7573f0955625aa96726f3494766172fb62fd2119cb6b3d933a27eedf3 not found: ID does not exist" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.726962 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gtmvb"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.730911 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gtmvb"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.733825 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vvxzs"] Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.734684 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:56:04 crc kubenswrapper[4715]: W1203 21:56:04.734708 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7912273_d7d2_4887_8525_e0c9e20d569a.slice/crio-d50f21600fffc73f942a6bcbe998e287429413a14f87de4ad3f53f5daf602217 WatchSource:0}: Error finding container d50f21600fffc73f942a6bcbe998e287429413a14f87de4ad3f53f5daf602217: Status 404 returned error can't find the container with id d50f21600fffc73f942a6bcbe998e287429413a14f87de4ad3f53f5daf602217 Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.772385 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-operator-metrics\") pod \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.772424 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcb4b\" (UniqueName: \"kubernetes.io/projected/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-kube-api-access-tcb4b\") pod \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.772496 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-trusted-ca\") pod \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\" (UID: \"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8\") " Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.775111 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" (UID: "ff7576eb-0d55-4e08-9ee8-734dad0aa0d8"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.778492 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-kube-api-access-tcb4b" (OuterVolumeSpecName: "kube-api-access-tcb4b") pod "ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" (UID: "ff7576eb-0d55-4e08-9ee8-734dad0aa0d8"). InnerVolumeSpecName "kube-api-access-tcb4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.779515 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" (UID: "ff7576eb-0d55-4e08-9ee8-734dad0aa0d8"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.791240 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l88hv"] Dec 03 21:56:04 crc kubenswrapper[4715]: W1203 21:56:04.804308 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3eda92a8_2005_4801_befd_d1db15f1b2fa.slice/crio-c958c2298e780570afd70ae28aa30418386f67d58145948a562e03768d0889b0 WatchSource:0}: Error finding container c958c2298e780570afd70ae28aa30418386f67d58145948a562e03768d0889b0: Status 404 returned error can't find the container with id c958c2298e780570afd70ae28aa30418386f67d58145948a562e03768d0889b0 Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.873972 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.874005 4715 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:04 crc kubenswrapper[4715]: I1203 21:56:04.874015 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcb4b\" (UniqueName: \"kubernetes.io/projected/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8-kube-api-access-tcb4b\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.370314 4715 generic.go:334] "Generic (PLEG): container finished" podID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerID="c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b" exitCode=0 Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.370449 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l88hv" event={"ID":"3eda92a8-2005-4801-befd-d1db15f1b2fa","Type":"ContainerDied","Data":"c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b"} Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.370987 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l88hv" event={"ID":"3eda92a8-2005-4801-befd-d1db15f1b2fa","Type":"ContainerStarted","Data":"c958c2298e780570afd70ae28aa30418386f67d58145948a562e03768d0889b0"} Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.374725 4715 generic.go:334] "Generic (PLEG): container finished" podID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerID="56126bbb4ead20dc7edde612444dae2c4b37e4c1cc86fbff07b6c4c979565ca9" exitCode=0 Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.375062 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvxzs" event={"ID":"f7912273-d7d2-4887-8525-e0c9e20d569a","Type":"ContainerDied","Data":"56126bbb4ead20dc7edde612444dae2c4b37e4c1cc86fbff07b6c4c979565ca9"} Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.375116 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvxzs" event={"ID":"f7912273-d7d2-4887-8525-e0c9e20d569a","Type":"ContainerStarted","Data":"d50f21600fffc73f942a6bcbe998e287429413a14f87de4ad3f53f5daf602217"} Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.375662 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.384664 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" event={"ID":"62dffbaa-068c-484d-822d-51bb3bfd3105","Type":"ContainerStarted","Data":"32f326bf13bdc173a1bfd4daeacebe4959c8f1ba5e31f490a3090663e2a23e4c"} Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.384767 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" event={"ID":"62dffbaa-068c-484d-822d-51bb3bfd3105","Type":"ContainerStarted","Data":"eb7b4a5f6d5928e38fb1474de355c01b3a3efb5616fcaf0a407b22f0e84e2223"} Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.387122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.393383 4715 generic.go:334] "Generic (PLEG): container finished" podID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerID="f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd" exitCode=0 Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.393485 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" event={"ID":"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8","Type":"ContainerDied","Data":"f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd"} Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.393581 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" event={"ID":"ff7576eb-0d55-4e08-9ee8-734dad0aa0d8","Type":"ContainerDied","Data":"a4c15e9c44b258184e9f8eaab2661f0542ad09b29ced34d739a544cc560cde0e"} Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.393626 4715 scope.go:117] "RemoveContainer" containerID="f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.393752 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-28k2p" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.394570 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.418049 4715 scope.go:117] "RemoveContainer" containerID="fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.433888 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-28k2p"] Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.437697 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-28k2p"] Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.448734 4715 scope.go:117] "RemoveContainer" containerID="f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd" Dec 03 21:56:05 crc kubenswrapper[4715]: E1203 21:56:05.449415 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd\": container with ID starting with f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd not found: ID does not exist" containerID="f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.449496 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd"} err="failed to get container status \"f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd\": rpc error: code = NotFound desc = could not find container \"f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd\": container with ID starting with f2d62b4999029cc7156294247363be06564bb8bc4fa16e74ec14d24d314332bd not found: ID does not exist" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.449754 4715 scope.go:117] "RemoveContainer" containerID="fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab" Dec 03 21:56:05 crc kubenswrapper[4715]: E1203 21:56:05.450235 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab\": container with ID starting with fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab not found: ID does not exist" containerID="fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.450268 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab"} err="failed to get container status \"fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab\": rpc error: code = NotFound desc = could not find container \"fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab\": container with ID starting with fc3ec6c02e340c951d49812747dcaa6551651f9d4032a4e529a7c219394ea2ab not found: ID does not exist" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.476332 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-ss8p4" podStartSLOduration=2.47629786 podStartE2EDuration="2.47629786s" podCreationTimestamp="2025-12-03 21:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:56:05.456412253 +0000 UTC m=+862.199122918" watchObservedRunningTime="2025-12-03 21:56:05.47629786 +0000 UTC m=+862.219008495" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.647698 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" path="/var/lib/kubelet/pods/2303c745-dd7b-4ccc-b0f8-7cb25c6efea5/volumes" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.650241 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" path="/var/lib/kubelet/pods/28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e/volumes" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.651621 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d631c483-9492-4253-8029-078d1120650d" path="/var/lib/kubelet/pods/d631c483-9492-4253-8029-078d1120650d/volumes" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.654275 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" path="/var/lib/kubelet/pods/ff7576eb-0d55-4e08-9ee8-734dad0aa0d8/volumes" Dec 03 21:56:05 crc kubenswrapper[4715]: I1203 21:56:05.655996 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" path="/var/lib/kubelet/pods/ffc954fc-ce5b-48c0-89ba-ebca500a7eee/volumes" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.371677 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-v8gpf"] Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.371980 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d631c483-9492-4253-8029-078d1120650d" containerName="extract-content" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372002 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d631c483-9492-4253-8029-078d1120650d" containerName="extract-content" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372024 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerName="extract-utilities" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372037 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerName="extract-utilities" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372054 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372066 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372085 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372097 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372112 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="extract-content" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372124 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="extract-content" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372139 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372152 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372165 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d631c483-9492-4253-8029-078d1120650d" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372177 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d631c483-9492-4253-8029-078d1120650d" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372192 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="extract-utilities" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372206 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="extract-utilities" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372225 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerName="extract-utilities" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372238 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerName="extract-utilities" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372257 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372272 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372290 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d631c483-9492-4253-8029-078d1120650d" containerName="extract-utilities" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372304 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d631c483-9492-4253-8029-078d1120650d" containerName="extract-utilities" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372325 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerName="extract-content" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372392 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerName="extract-content" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372413 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372425 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" Dec 03 21:56:06 crc kubenswrapper[4715]: E1203 21:56:06.372445 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerName="extract-content" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372458 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerName="extract-content" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372645 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d631c483-9492-4253-8029-078d1120650d" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372669 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372684 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2303c745-dd7b-4ccc-b0f8-7cb25c6efea5" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372698 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff7576eb-0d55-4e08-9ee8-734dad0aa0d8" containerName="marketplace-operator" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372720 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffc954fc-ce5b-48c0-89ba-ebca500a7eee" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.372737 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="28ef7bf3-ca1d-44d8-b647-60fd71fe9c3e" containerName="registry-server" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.374285 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.378075 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.392597 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v8gpf"] Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.404348 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554ea49c-3cc3-42fd-b080-3054daed5166-utilities\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.404486 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6thjq\" (UniqueName: \"kubernetes.io/projected/554ea49c-3cc3-42fd-b080-3054daed5166-kube-api-access-6thjq\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.404590 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554ea49c-3cc3-42fd-b080-3054daed5166-catalog-content\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.505061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6thjq\" (UniqueName: \"kubernetes.io/projected/554ea49c-3cc3-42fd-b080-3054daed5166-kube-api-access-6thjq\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.505272 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554ea49c-3cc3-42fd-b080-3054daed5166-catalog-content\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.505419 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554ea49c-3cc3-42fd-b080-3054daed5166-utilities\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.506107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/554ea49c-3cc3-42fd-b080-3054daed5166-utilities\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.507581 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/554ea49c-3cc3-42fd-b080-3054daed5166-catalog-content\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.535689 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6thjq\" (UniqueName: \"kubernetes.io/projected/554ea49c-3cc3-42fd-b080-3054daed5166-kube-api-access-6thjq\") pod \"redhat-marketplace-v8gpf\" (UID: \"554ea49c-3cc3-42fd-b080-3054daed5166\") " pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.560304 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4g7ql"] Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.561586 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.566228 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.579156 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4g7ql"] Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.606254 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9l5j\" (UniqueName: \"kubernetes.io/projected/c759bed5-8b17-4009-90f1-6422be78653d-kube-api-access-f9l5j\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.606321 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c759bed5-8b17-4009-90f1-6422be78653d-utilities\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.606340 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c759bed5-8b17-4009-90f1-6422be78653d-catalog-content\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.701572 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.708255 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9l5j\" (UniqueName: \"kubernetes.io/projected/c759bed5-8b17-4009-90f1-6422be78653d-kube-api-access-f9l5j\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.708342 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c759bed5-8b17-4009-90f1-6422be78653d-utilities\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.708372 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c759bed5-8b17-4009-90f1-6422be78653d-catalog-content\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.709061 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c759bed5-8b17-4009-90f1-6422be78653d-catalog-content\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.709594 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c759bed5-8b17-4009-90f1-6422be78653d-utilities\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.732209 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9l5j\" (UniqueName: \"kubernetes.io/projected/c759bed5-8b17-4009-90f1-6422be78653d-kube-api-access-f9l5j\") pod \"redhat-operators-4g7ql\" (UID: \"c759bed5-8b17-4009-90f1-6422be78653d\") " pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.763044 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bx8qp"] Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.764047 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.781158 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx8qp"] Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.809352 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-utilities\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.809432 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nx8l\" (UniqueName: \"kubernetes.io/projected/d0bce231-d2cd-44fe-b55e-ee830798ad3f-kube-api-access-8nx8l\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.809531 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-catalog-content\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.911394 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-utilities\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.911441 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nx8l\" (UniqueName: \"kubernetes.io/projected/d0bce231-d2cd-44fe-b55e-ee830798ad3f-kube-api-access-8nx8l\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.911489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-catalog-content\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.912453 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-catalog-content\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.912972 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-utilities\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.917906 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.934130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nx8l\" (UniqueName: \"kubernetes.io/projected/d0bce231-d2cd-44fe-b55e-ee830798ad3f-kube-api-access-8nx8l\") pod \"redhat-marketplace-bx8qp\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.943045 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-v8gpf"] Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.965710 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vj7mq"] Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.966869 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:06 crc kubenswrapper[4715]: I1203 21:56:06.969370 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vj7mq"] Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.013297 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-utilities\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.013366 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-catalog-content\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.013402 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78ndm\" (UniqueName: \"kubernetes.io/projected/4629f441-9cf2-469c-9815-5c1739cc429a-kube-api-access-78ndm\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.091169 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.115020 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-utilities\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.115096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-catalog-content\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.115132 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78ndm\" (UniqueName: \"kubernetes.io/projected/4629f441-9cf2-469c-9815-5c1739cc429a-kube-api-access-78ndm\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.115681 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-utilities\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.116096 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-catalog-content\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.139245 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78ndm\" (UniqueName: \"kubernetes.io/projected/4629f441-9cf2-469c-9815-5c1739cc429a-kube-api-access-78ndm\") pod \"redhat-operators-vj7mq\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.170250 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4g7ql"] Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.376829 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.438090 4715 generic.go:334] "Generic (PLEG): container finished" podID="c759bed5-8b17-4009-90f1-6422be78653d" containerID="87881bec2024db002083ab852d5fef60f081bf1aaec8da0c32b5ca517707f2c6" exitCode=0 Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.438170 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4g7ql" event={"ID":"c759bed5-8b17-4009-90f1-6422be78653d","Type":"ContainerDied","Data":"87881bec2024db002083ab852d5fef60f081bf1aaec8da0c32b5ca517707f2c6"} Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.438713 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4g7ql" event={"ID":"c759bed5-8b17-4009-90f1-6422be78653d","Type":"ContainerStarted","Data":"52083e7b0696576f97f5a32219c942ab28ca544c8c8a9c411b2174c1bf482cb5"} Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.450711 4715 generic.go:334] "Generic (PLEG): container finished" podID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerID="92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb" exitCode=0 Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.450796 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l88hv" event={"ID":"3eda92a8-2005-4801-befd-d1db15f1b2fa","Type":"ContainerDied","Data":"92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb"} Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.455265 4715 generic.go:334] "Generic (PLEG): container finished" podID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerID="25cc918553bafd1a7bd4a4878cddc9488ac37b227994a1de2cd4e938c8d9333d" exitCode=0 Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.455391 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvxzs" event={"ID":"f7912273-d7d2-4887-8525-e0c9e20d569a","Type":"ContainerDied","Data":"25cc918553bafd1a7bd4a4878cddc9488ac37b227994a1de2cd4e938c8d9333d"} Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.474970 4715 generic.go:334] "Generic (PLEG): container finished" podID="554ea49c-3cc3-42fd-b080-3054daed5166" containerID="d26d5ff4fd578223c793292f220fb754177d7e0b94d8b1b79bc9704de8cfd015" exitCode=0 Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.475143 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8gpf" event={"ID":"554ea49c-3cc3-42fd-b080-3054daed5166","Type":"ContainerDied","Data":"d26d5ff4fd578223c793292f220fb754177d7e0b94d8b1b79bc9704de8cfd015"} Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.475222 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8gpf" event={"ID":"554ea49c-3cc3-42fd-b080-3054daed5166","Type":"ContainerStarted","Data":"18182210cbf33c6a36f69fd17780bfb8c1374d3aafdda56c28c2d70c8af98948"} Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.570635 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx8qp"] Dec 03 21:56:07 crc kubenswrapper[4715]: I1203 21:56:07.848777 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vj7mq"] Dec 03 21:56:07 crc kubenswrapper[4715]: W1203 21:56:07.857008 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4629f441_9cf2_469c_9815_5c1739cc429a.slice/crio-61d2221d7874cbf20f6bb5f21c949b4b81d0303e48742f55c06d121b47ac5f31 WatchSource:0}: Error finding container 61d2221d7874cbf20f6bb5f21c949b4b81d0303e48742f55c06d121b47ac5f31: Status 404 returned error can't find the container with id 61d2221d7874cbf20f6bb5f21c949b4b81d0303e48742f55c06d121b47ac5f31 Dec 03 21:56:08 crc kubenswrapper[4715]: I1203 21:56:08.485810 4715 generic.go:334] "Generic (PLEG): container finished" podID="4629f441-9cf2-469c-9815-5c1739cc429a" containerID="c0f8ff110ecf8500b867b55c5ac32d8dcc4dba1982bf90340b67edf5bda7107d" exitCode=0 Dec 03 21:56:08 crc kubenswrapper[4715]: I1203 21:56:08.485882 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj7mq" event={"ID":"4629f441-9cf2-469c-9815-5c1739cc429a","Type":"ContainerDied","Data":"c0f8ff110ecf8500b867b55c5ac32d8dcc4dba1982bf90340b67edf5bda7107d"} Dec 03 21:56:08 crc kubenswrapper[4715]: I1203 21:56:08.486413 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj7mq" event={"ID":"4629f441-9cf2-469c-9815-5c1739cc429a","Type":"ContainerStarted","Data":"61d2221d7874cbf20f6bb5f21c949b4b81d0303e48742f55c06d121b47ac5f31"} Dec 03 21:56:08 crc kubenswrapper[4715]: I1203 21:56:08.489181 4715 generic.go:334] "Generic (PLEG): container finished" podID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerID="ccfbb72d50a27cc4a75ba0a1ee5bb2644fff3d31afac00d00137a464a56dfcb0" exitCode=0 Dec 03 21:56:08 crc kubenswrapper[4715]: I1203 21:56:08.489267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx8qp" event={"ID":"d0bce231-d2cd-44fe-b55e-ee830798ad3f","Type":"ContainerDied","Data":"ccfbb72d50a27cc4a75ba0a1ee5bb2644fff3d31afac00d00137a464a56dfcb0"} Dec 03 21:56:08 crc kubenswrapper[4715]: I1203 21:56:08.489299 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx8qp" event={"ID":"d0bce231-d2cd-44fe-b55e-ee830798ad3f","Type":"ContainerStarted","Data":"a771ffe9ef97739b13c343e66e55601d1eee0405dec854be1ce322a259d27a5d"} Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.179073 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9nxq5"] Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.180303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.192308 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9nxq5"] Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.277061 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2c2g9\" (UniqueName: \"kubernetes.io/projected/2d979921-24da-4ea9-bb4a-bb7f921778e0-kube-api-access-2c2g9\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.277125 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d979921-24da-4ea9-bb4a-bb7f921778e0-catalog-content\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.277249 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d979921-24da-4ea9-bb4a-bb7f921778e0-utilities\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.360900 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rvbck"] Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.362549 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.374916 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rvbck"] Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.377894 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d979921-24da-4ea9-bb4a-bb7f921778e0-utilities\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.377945 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9w2d\" (UniqueName: \"kubernetes.io/projected/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-kube-api-access-m9w2d\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.377968 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-utilities\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.377991 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-catalog-content\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.378019 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2c2g9\" (UniqueName: \"kubernetes.io/projected/2d979921-24da-4ea9-bb4a-bb7f921778e0-kube-api-access-2c2g9\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.378189 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d979921-24da-4ea9-bb4a-bb7f921778e0-catalog-content\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.378544 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d979921-24da-4ea9-bb4a-bb7f921778e0-utilities\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.378588 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d979921-24da-4ea9-bb4a-bb7f921778e0-catalog-content\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.402561 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2c2g9\" (UniqueName: \"kubernetes.io/projected/2d979921-24da-4ea9-bb4a-bb7f921778e0-kube-api-access-2c2g9\") pod \"community-operators-9nxq5\" (UID: \"2d979921-24da-4ea9-bb4a-bb7f921778e0\") " pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.479835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9w2d\" (UniqueName: \"kubernetes.io/projected/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-kube-api-access-m9w2d\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.479905 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-utilities\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.479942 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-catalog-content\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.480719 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-catalog-content\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.500569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-utilities\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.502901 4715 generic.go:334] "Generic (PLEG): container finished" podID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerID="659b9568b2c63ffc1e4a98199293e6c965fe03713f70cd1bfc67efcb91c6be42" exitCode=0 Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.503018 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx8qp" event={"ID":"d0bce231-d2cd-44fe-b55e-ee830798ad3f","Type":"ContainerDied","Data":"659b9568b2c63ffc1e4a98199293e6c965fe03713f70cd1bfc67efcb91c6be42"} Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.509271 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9w2d\" (UniqueName: \"kubernetes.io/projected/63ed1f29-d7e8-4dc2-b660-dbac35ed94ec-kube-api-access-m9w2d\") pod \"certified-operators-rvbck\" (UID: \"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec\") " pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.510436 4715 generic.go:334] "Generic (PLEG): container finished" podID="554ea49c-3cc3-42fd-b080-3054daed5166" containerID="05439845f7e124f32d3bcdf648d66901452b4d27e92416fc21014877faf2d33b" exitCode=0 Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.510601 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8gpf" event={"ID":"554ea49c-3cc3-42fd-b080-3054daed5166","Type":"ContainerDied","Data":"05439845f7e124f32d3bcdf648d66901452b4d27e92416fc21014877faf2d33b"} Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.519251 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj7mq" event={"ID":"4629f441-9cf2-469c-9815-5c1739cc429a","Type":"ContainerStarted","Data":"6fdb12365d63bffca143f2580ad96b2cae63438d750c118301dee52d1a07d53d"} Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.541918 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.550931 4715 generic.go:334] "Generic (PLEG): container finished" podID="c759bed5-8b17-4009-90f1-6422be78653d" containerID="485ba0dbbf5751991f20ecbd94bf3d8a31c4a3da17e7d9c568fe9a124caa26d1" exitCode=0 Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.551022 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4g7ql" event={"ID":"c759bed5-8b17-4009-90f1-6422be78653d","Type":"ContainerDied","Data":"485ba0dbbf5751991f20ecbd94bf3d8a31c4a3da17e7d9c568fe9a124caa26d1"} Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.576953 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l88hv" event={"ID":"3eda92a8-2005-4801-befd-d1db15f1b2fa","Type":"ContainerStarted","Data":"517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965"} Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.579330 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvxzs" event={"ID":"f7912273-d7d2-4887-8525-e0c9e20d569a","Type":"ContainerStarted","Data":"3bd538cdd772a41e40a9d772a8f6cb0d30c3c2f5a1dc36af515b3d75b13ba248"} Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.649829 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l88hv" podStartSLOduration=2.5630896869999997 podStartE2EDuration="5.649813444s" podCreationTimestamp="2025-12-03 21:56:04 +0000 UTC" firstStartedPulling="2025-12-03 21:56:05.375185885 +0000 UTC m=+862.117896520" lastFinishedPulling="2025-12-03 21:56:08.461909652 +0000 UTC m=+865.204620277" observedRunningTime="2025-12-03 21:56:09.649163297 +0000 UTC m=+866.391873892" watchObservedRunningTime="2025-12-03 21:56:09.649813444 +0000 UTC m=+866.392524029" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.710342 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.731297 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vvxzs" podStartSLOduration=3.764305388 podStartE2EDuration="6.731271609s" podCreationTimestamp="2025-12-03 21:56:03 +0000 UTC" firstStartedPulling="2025-12-03 21:56:05.381101995 +0000 UTC m=+862.123812620" lastFinishedPulling="2025-12-03 21:56:08.348068246 +0000 UTC m=+865.090778841" observedRunningTime="2025-12-03 21:56:09.719529112 +0000 UTC m=+866.462239707" watchObservedRunningTime="2025-12-03 21:56:09.731271609 +0000 UTC m=+866.473982204" Dec 03 21:56:09 crc kubenswrapper[4715]: I1203 21:56:09.930094 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rvbck"] Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.058449 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9nxq5"] Dec 03 21:56:10 crc kubenswrapper[4715]: W1203 21:56:10.105356 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d979921_24da_4ea9_bb4a_bb7f921778e0.slice/crio-c07ca9d261eb09f24905eb5689ebaa9c38273cfd5ed72055cb5a06c6928b0131 WatchSource:0}: Error finding container c07ca9d261eb09f24905eb5689ebaa9c38273cfd5ed72055cb5a06c6928b0131: Status 404 returned error can't find the container with id c07ca9d261eb09f24905eb5689ebaa9c38273cfd5ed72055cb5a06c6928b0131 Dec 03 21:56:10 crc kubenswrapper[4715]: W1203 21:56:10.113014 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63ed1f29_d7e8_4dc2_b660_dbac35ed94ec.slice/crio-ab401459a7f76ef2ed046de9c96257a1382c127ccfeb4008369a787e9cd6c0f5 WatchSource:0}: Error finding container ab401459a7f76ef2ed046de9c96257a1382c127ccfeb4008369a787e9cd6c0f5: Status 404 returned error can't find the container with id ab401459a7f76ef2ed046de9c96257a1382c127ccfeb4008369a787e9cd6c0f5 Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.585919 4715 generic.go:334] "Generic (PLEG): container finished" podID="63ed1f29-d7e8-4dc2-b660-dbac35ed94ec" containerID="86ae31b307058a758f46bb23d1b771708c6de62cdfee184f83d76b19795fb63c" exitCode=0 Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.586142 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvbck" event={"ID":"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec","Type":"ContainerDied","Data":"86ae31b307058a758f46bb23d1b771708c6de62cdfee184f83d76b19795fb63c"} Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.587368 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvbck" event={"ID":"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec","Type":"ContainerStarted","Data":"ab401459a7f76ef2ed046de9c96257a1382c127ccfeb4008369a787e9cd6c0f5"} Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.589937 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx8qp" event={"ID":"d0bce231-d2cd-44fe-b55e-ee830798ad3f","Type":"ContainerStarted","Data":"a738cb6adab6591958c9421852fb59ef293e9355099621e40e106866f90c7b8c"} Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.591728 4715 generic.go:334] "Generic (PLEG): container finished" podID="4629f441-9cf2-469c-9815-5c1739cc429a" containerID="6fdb12365d63bffca143f2580ad96b2cae63438d750c118301dee52d1a07d53d" exitCode=0 Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.591764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj7mq" event={"ID":"4629f441-9cf2-469c-9815-5c1739cc429a","Type":"ContainerDied","Data":"6fdb12365d63bffca143f2580ad96b2cae63438d750c118301dee52d1a07d53d"} Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.593539 4715 generic.go:334] "Generic (PLEG): container finished" podID="2d979921-24da-4ea9-bb4a-bb7f921778e0" containerID="0077c8d3b9653970918f87b1d56dfa6c968b0862928ac79a36fe7923c35a29a6" exitCode=0 Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.593607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nxq5" event={"ID":"2d979921-24da-4ea9-bb4a-bb7f921778e0","Type":"ContainerDied","Data":"0077c8d3b9653970918f87b1d56dfa6c968b0862928ac79a36fe7923c35a29a6"} Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.593634 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nxq5" event={"ID":"2d979921-24da-4ea9-bb4a-bb7f921778e0","Type":"ContainerStarted","Data":"c07ca9d261eb09f24905eb5689ebaa9c38273cfd5ed72055cb5a06c6928b0131"} Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.597107 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-v8gpf" event={"ID":"554ea49c-3cc3-42fd-b080-3054daed5166","Type":"ContainerStarted","Data":"a3182e6181028c9a4a5dffb6d2962c904a91f3257d3eb3655fc32754352f7e4a"} Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.603088 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4g7ql" event={"ID":"c759bed5-8b17-4009-90f1-6422be78653d","Type":"ContainerStarted","Data":"911631edb53a462877786c35673234b5f19f783f40d3b2b57a4192f723c1d249"} Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.634143 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-v8gpf" podStartSLOduration=1.946780344 podStartE2EDuration="4.634118231s" podCreationTimestamp="2025-12-03 21:56:06 +0000 UTC" firstStartedPulling="2025-12-03 21:56:07.478313854 +0000 UTC m=+864.221024459" lastFinishedPulling="2025-12-03 21:56:10.165651751 +0000 UTC m=+866.908362346" observedRunningTime="2025-12-03 21:56:10.633385022 +0000 UTC m=+867.376095627" watchObservedRunningTime="2025-12-03 21:56:10.634118231 +0000 UTC m=+867.376828826" Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.657730 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4g7ql" podStartSLOduration=1.858811474 podStartE2EDuration="4.657711187s" podCreationTimestamp="2025-12-03 21:56:06 +0000 UTC" firstStartedPulling="2025-12-03 21:56:07.440445853 +0000 UTC m=+864.183156448" lastFinishedPulling="2025-12-03 21:56:10.239345566 +0000 UTC m=+866.982056161" observedRunningTime="2025-12-03 21:56:10.657410269 +0000 UTC m=+867.400120874" watchObservedRunningTime="2025-12-03 21:56:10.657711187 +0000 UTC m=+867.400421792" Dec 03 21:56:10 crc kubenswrapper[4715]: I1203 21:56:10.683899 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bx8qp" podStartSLOduration=3.296157687 podStartE2EDuration="4.683870372s" podCreationTimestamp="2025-12-03 21:56:06 +0000 UTC" firstStartedPulling="2025-12-03 21:56:08.506921095 +0000 UTC m=+865.249631730" lastFinishedPulling="2025-12-03 21:56:09.89463382 +0000 UTC m=+866.637344415" observedRunningTime="2025-12-03 21:56:10.680930543 +0000 UTC m=+867.423641138" watchObservedRunningTime="2025-12-03 21:56:10.683870372 +0000 UTC m=+867.426580967" Dec 03 21:56:11 crc kubenswrapper[4715]: I1203 21:56:11.619692 4715 generic.go:334] "Generic (PLEG): container finished" podID="63ed1f29-d7e8-4dc2-b660-dbac35ed94ec" containerID="7abd7dd31b66e4f10bd86f38a6e7ce3a0b5aaddb6f98f82895d6e6df99902ea4" exitCode=0 Dec 03 21:56:11 crc kubenswrapper[4715]: I1203 21:56:11.619748 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvbck" event={"ID":"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec","Type":"ContainerDied","Data":"7abd7dd31b66e4f10bd86f38a6e7ce3a0b5aaddb6f98f82895d6e6df99902ea4"} Dec 03 21:56:11 crc kubenswrapper[4715]: I1203 21:56:11.624464 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj7mq" event={"ID":"4629f441-9cf2-469c-9815-5c1739cc429a","Type":"ContainerStarted","Data":"7b89894e4cbe371d2f861bb14b093b8c1d5ab07f33cca597b2adc25ecabecf2d"} Dec 03 21:56:11 crc kubenswrapper[4715]: I1203 21:56:11.631238 4715 generic.go:334] "Generic (PLEG): container finished" podID="2d979921-24da-4ea9-bb4a-bb7f921778e0" containerID="e2eaeb05b4bb4d040b3e67f57e53231268c5f2935d3c42a81462cbd228b433c6" exitCode=0 Dec 03 21:56:11 crc kubenswrapper[4715]: I1203 21:56:11.636599 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nxq5" event={"ID":"2d979921-24da-4ea9-bb4a-bb7f921778e0","Type":"ContainerDied","Data":"e2eaeb05b4bb4d040b3e67f57e53231268c5f2935d3c42a81462cbd228b433c6"} Dec 03 21:56:11 crc kubenswrapper[4715]: I1203 21:56:11.690679 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vj7mq" podStartSLOduration=3.21279364 podStartE2EDuration="5.690640694s" podCreationTimestamp="2025-12-03 21:56:06 +0000 UTC" firstStartedPulling="2025-12-03 21:56:08.508051995 +0000 UTC m=+865.250762620" lastFinishedPulling="2025-12-03 21:56:10.985899079 +0000 UTC m=+867.728609674" observedRunningTime="2025-12-03 21:56:11.685921448 +0000 UTC m=+868.428632053" watchObservedRunningTime="2025-12-03 21:56:11.690640694 +0000 UTC m=+868.433351279" Dec 03 21:56:13 crc kubenswrapper[4715]: I1203 21:56:13.645745 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvbck" event={"ID":"63ed1f29-d7e8-4dc2-b660-dbac35ed94ec","Type":"ContainerStarted","Data":"1346cc354b769b1545161a185145d6c3fe7599a750efe541a01c18e3feb83e9d"} Dec 03 21:56:13 crc kubenswrapper[4715]: I1203 21:56:13.647492 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9nxq5" event={"ID":"2d979921-24da-4ea9-bb4a-bb7f921778e0","Type":"ContainerStarted","Data":"9826f24bf1219b26203428d69385c38103dd0c733288a99a6303c0fe01a7d13d"} Dec 03 21:56:13 crc kubenswrapper[4715]: I1203 21:56:13.674046 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rvbck" podStartSLOduration=2.608489322 podStartE2EDuration="4.674016257s" podCreationTimestamp="2025-12-03 21:56:09 +0000 UTC" firstStartedPulling="2025-12-03 21:56:10.588301448 +0000 UTC m=+867.331012043" lastFinishedPulling="2025-12-03 21:56:12.653828373 +0000 UTC m=+869.396538978" observedRunningTime="2025-12-03 21:56:13.669883076 +0000 UTC m=+870.412593671" watchObservedRunningTime="2025-12-03 21:56:13.674016257 +0000 UTC m=+870.416726872" Dec 03 21:56:13 crc kubenswrapper[4715]: I1203 21:56:13.698090 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9nxq5" podStartSLOduration=2.55089922 podStartE2EDuration="4.698070285s" podCreationTimestamp="2025-12-03 21:56:09 +0000 UTC" firstStartedPulling="2025-12-03 21:56:10.595265525 +0000 UTC m=+867.337976120" lastFinishedPulling="2025-12-03 21:56:12.74243655 +0000 UTC m=+869.485147185" observedRunningTime="2025-12-03 21:56:13.695149277 +0000 UTC m=+870.437859872" watchObservedRunningTime="2025-12-03 21:56:13.698070285 +0000 UTC m=+870.440780880" Dec 03 21:56:14 crc kubenswrapper[4715]: I1203 21:56:14.228833 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:14 crc kubenswrapper[4715]: I1203 21:56:14.229427 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:14 crc kubenswrapper[4715]: I1203 21:56:14.280573 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:14 crc kubenswrapper[4715]: I1203 21:56:14.545013 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:14 crc kubenswrapper[4715]: I1203 21:56:14.545079 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:14 crc kubenswrapper[4715]: I1203 21:56:14.587679 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:14 crc kubenswrapper[4715]: I1203 21:56:14.699147 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:14 crc kubenswrapper[4715]: I1203 21:56:14.701108 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:16 crc kubenswrapper[4715]: I1203 21:56:16.701745 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:16 crc kubenswrapper[4715]: I1203 21:56:16.701815 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:16 crc kubenswrapper[4715]: I1203 21:56:16.749243 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:16 crc kubenswrapper[4715]: I1203 21:56:16.918472 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:16 crc kubenswrapper[4715]: I1203 21:56:16.918655 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:16 crc kubenswrapper[4715]: I1203 21:56:16.969916 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.091941 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.092365 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.149909 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.377921 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.378045 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.431021 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.553875 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vvxzs"] Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.554174 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vvxzs" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerName="registry-server" containerID="cri-o://3bd538cdd772a41e40a9d772a8f6cb0d30c3c2f5a1dc36af515b3d75b13ba248" gracePeriod=2 Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.732834 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4g7ql" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.738247 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.745764 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.746981 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-v8gpf" Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.757178 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l88hv"] Dec 03 21:56:17 crc kubenswrapper[4715]: I1203 21:56:17.757610 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-l88hv" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerName="registry-server" containerID="cri-o://517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965" gracePeriod=2 Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.486958 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.543798 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.543917 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.603804 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.641975 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-utilities\") pod \"3eda92a8-2005-4801-befd-d1db15f1b2fa\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.643523 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-utilities" (OuterVolumeSpecName: "utilities") pod "3eda92a8-2005-4801-befd-d1db15f1b2fa" (UID: "3eda92a8-2005-4801-befd-d1db15f1b2fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.644448 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-catalog-content\") pod \"3eda92a8-2005-4801-befd-d1db15f1b2fa\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.644542 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5cj5\" (UniqueName: \"kubernetes.io/projected/3eda92a8-2005-4801-befd-d1db15f1b2fa-kube-api-access-n5cj5\") pod \"3eda92a8-2005-4801-befd-d1db15f1b2fa\" (UID: \"3eda92a8-2005-4801-befd-d1db15f1b2fa\") " Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.644750 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.656771 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eda92a8-2005-4801-befd-d1db15f1b2fa-kube-api-access-n5cj5" (OuterVolumeSpecName: "kube-api-access-n5cj5") pod "3eda92a8-2005-4801-befd-d1db15f1b2fa" (UID: "3eda92a8-2005-4801-befd-d1db15f1b2fa"). InnerVolumeSpecName "kube-api-access-n5cj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.695951 4715 generic.go:334] "Generic (PLEG): container finished" podID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerID="517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965" exitCode=0 Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.696035 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l88hv" event={"ID":"3eda92a8-2005-4801-befd-d1db15f1b2fa","Type":"ContainerDied","Data":"517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965"} Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.696134 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l88hv" event={"ID":"3eda92a8-2005-4801-befd-d1db15f1b2fa","Type":"ContainerDied","Data":"c958c2298e780570afd70ae28aa30418386f67d58145948a562e03768d0889b0"} Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.696154 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l88hv" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.696206 4715 scope.go:117] "RemoveContainer" containerID="517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.700031 4715 generic.go:334] "Generic (PLEG): container finished" podID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerID="3bd538cdd772a41e40a9d772a8f6cb0d30c3c2f5a1dc36af515b3d75b13ba248" exitCode=0 Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.700115 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvxzs" event={"ID":"f7912273-d7d2-4887-8525-e0c9e20d569a","Type":"ContainerDied","Data":"3bd538cdd772a41e40a9d772a8f6cb0d30c3c2f5a1dc36af515b3d75b13ba248"} Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.701459 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3eda92a8-2005-4801-befd-d1db15f1b2fa" (UID: "3eda92a8-2005-4801-befd-d1db15f1b2fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.714672 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.714770 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.731697 4715 scope.go:117] "RemoveContainer" containerID="92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.748189 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5cj5\" (UniqueName: \"kubernetes.io/projected/3eda92a8-2005-4801-befd-d1db15f1b2fa-kube-api-access-n5cj5\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.748261 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3eda92a8-2005-4801-befd-d1db15f1b2fa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.756005 4715 scope.go:117] "RemoveContainer" containerID="c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.760968 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9nxq5" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.777427 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.796642 4715 scope.go:117] "RemoveContainer" containerID="517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965" Dec 03 21:56:19 crc kubenswrapper[4715]: E1203 21:56:19.798005 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965\": container with ID starting with 517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965 not found: ID does not exist" containerID="517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.798054 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965"} err="failed to get container status \"517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965\": rpc error: code = NotFound desc = could not find container \"517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965\": container with ID starting with 517958103db6c058fe604918a95d207f0c9fa36e9c0426a525cf85beb096e965 not found: ID does not exist" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.798099 4715 scope.go:117] "RemoveContainer" containerID="92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb" Dec 03 21:56:19 crc kubenswrapper[4715]: E1203 21:56:19.798669 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb\": container with ID starting with 92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb not found: ID does not exist" containerID="92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.798795 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb"} err="failed to get container status \"92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb\": rpc error: code = NotFound desc = could not find container \"92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb\": container with ID starting with 92b6de45c5cc823a1f9268a5b2f3208aaad89d907ce1f291aa094aa8ef8b22fb not found: ID does not exist" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.798845 4715 scope.go:117] "RemoveContainer" containerID="c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b" Dec 03 21:56:19 crc kubenswrapper[4715]: E1203 21:56:19.799395 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b\": container with ID starting with c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b not found: ID does not exist" containerID="c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.799441 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b"} err="failed to get container status \"c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b\": rpc error: code = NotFound desc = could not find container \"c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b\": container with ID starting with c0357c693bae0bcf5f36779eda6703475e4532ffc1b041b3448033730f38c95b not found: ID does not exist" Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.952548 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx8qp"] Dec 03 21:56:19 crc kubenswrapper[4715]: I1203 21:56:19.953255 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bx8qp" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerName="registry-server" containerID="cri-o://a738cb6adab6591958c9421852fb59ef293e9355099621e40e106866f90c7b8c" gracePeriod=2 Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.024836 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-l88hv"] Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.028994 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-l88hv"] Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.154868 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vj7mq"] Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.155125 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vj7mq" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" containerName="registry-server" containerID="cri-o://7b89894e4cbe371d2f861bb14b093b8c1d5ab07f33cca597b2adc25ecabecf2d" gracePeriod=2 Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.641323 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.670855 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wm65\" (UniqueName: \"kubernetes.io/projected/f7912273-d7d2-4887-8525-e0c9e20d569a-kube-api-access-5wm65\") pod \"f7912273-d7d2-4887-8525-e0c9e20d569a\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.670989 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-catalog-content\") pod \"f7912273-d7d2-4887-8525-e0c9e20d569a\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.671041 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-utilities\") pod \"f7912273-d7d2-4887-8525-e0c9e20d569a\" (UID: \"f7912273-d7d2-4887-8525-e0c9e20d569a\") " Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.672079 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-utilities" (OuterVolumeSpecName: "utilities") pod "f7912273-d7d2-4887-8525-e0c9e20d569a" (UID: "f7912273-d7d2-4887-8525-e0c9e20d569a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.683900 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7912273-d7d2-4887-8525-e0c9e20d569a-kube-api-access-5wm65" (OuterVolumeSpecName: "kube-api-access-5wm65") pod "f7912273-d7d2-4887-8525-e0c9e20d569a" (UID: "f7912273-d7d2-4887-8525-e0c9e20d569a"). InnerVolumeSpecName "kube-api-access-5wm65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.709641 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vvxzs" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.709722 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vvxzs" event={"ID":"f7912273-d7d2-4887-8525-e0c9e20d569a","Type":"ContainerDied","Data":"d50f21600fffc73f942a6bcbe998e287429413a14f87de4ad3f53f5daf602217"} Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.709804 4715 scope.go:117] "RemoveContainer" containerID="3bd538cdd772a41e40a9d772a8f6cb0d30c3c2f5a1dc36af515b3d75b13ba248" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.729555 4715 scope.go:117] "RemoveContainer" containerID="25cc918553bafd1a7bd4a4878cddc9488ac37b227994a1de2cd4e938c8d9333d" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.742261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7912273-d7d2-4887-8525-e0c9e20d569a" (UID: "f7912273-d7d2-4887-8525-e0c9e20d569a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.748188 4715 scope.go:117] "RemoveContainer" containerID="56126bbb4ead20dc7edde612444dae2c4b37e4c1cc86fbff07b6c4c979565ca9" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.751673 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rvbck" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.772708 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.772760 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7912273-d7d2-4887-8525-e0c9e20d569a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:20 crc kubenswrapper[4715]: I1203 21:56:20.772773 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wm65\" (UniqueName: \"kubernetes.io/projected/f7912273-d7d2-4887-8525-e0c9e20d569a-kube-api-access-5wm65\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:21 crc kubenswrapper[4715]: I1203 21:56:21.056627 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vvxzs"] Dec 03 21:56:21 crc kubenswrapper[4715]: I1203 21:56:21.063665 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vvxzs"] Dec 03 21:56:21 crc kubenswrapper[4715]: I1203 21:56:21.641151 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" path="/var/lib/kubelet/pods/3eda92a8-2005-4801-befd-d1db15f1b2fa/volumes" Dec 03 21:56:21 crc kubenswrapper[4715]: I1203 21:56:21.641765 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" path="/var/lib/kubelet/pods/f7912273-d7d2-4887-8525-e0c9e20d569a/volumes" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:23.733749 4715 generic.go:334] "Generic (PLEG): container finished" podID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerID="a738cb6adab6591958c9421852fb59ef293e9355099621e40e106866f90c7b8c" exitCode=0 Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:23.733836 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx8qp" event={"ID":"d0bce231-d2cd-44fe-b55e-ee830798ad3f","Type":"ContainerDied","Data":"a738cb6adab6591958c9421852fb59ef293e9355099621e40e106866f90c7b8c"} Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:23.737018 4715 generic.go:334] "Generic (PLEG): container finished" podID="4629f441-9cf2-469c-9815-5c1739cc429a" containerID="7b89894e4cbe371d2f861bb14b093b8c1d5ab07f33cca597b2adc25ecabecf2d" exitCode=0 Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:23.737046 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj7mq" event={"ID":"4629f441-9cf2-469c-9815-5c1739cc429a","Type":"ContainerDied","Data":"7b89894e4cbe371d2f861bb14b093b8c1d5ab07f33cca597b2adc25ecabecf2d"} Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.587477 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.639921 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-utilities\") pod \"4629f441-9cf2-469c-9815-5c1739cc429a\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.640064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78ndm\" (UniqueName: \"kubernetes.io/projected/4629f441-9cf2-469c-9815-5c1739cc429a-kube-api-access-78ndm\") pod \"4629f441-9cf2-469c-9815-5c1739cc429a\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.640241 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-catalog-content\") pod \"4629f441-9cf2-469c-9815-5c1739cc429a\" (UID: \"4629f441-9cf2-469c-9815-5c1739cc429a\") " Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.640780 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-utilities" (OuterVolumeSpecName: "utilities") pod "4629f441-9cf2-469c-9815-5c1739cc429a" (UID: "4629f441-9cf2-469c-9815-5c1739cc429a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.650055 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4629f441-9cf2-469c-9815-5c1739cc429a-kube-api-access-78ndm" (OuterVolumeSpecName: "kube-api-access-78ndm") pod "4629f441-9cf2-469c-9815-5c1739cc429a" (UID: "4629f441-9cf2-469c-9815-5c1739cc429a"). InnerVolumeSpecName "kube-api-access-78ndm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.742664 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.742725 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78ndm\" (UniqueName: \"kubernetes.io/projected/4629f441-9cf2-469c-9815-5c1739cc429a-kube-api-access-78ndm\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.761391 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vj7mq" event={"ID":"4629f441-9cf2-469c-9815-5c1739cc429a","Type":"ContainerDied","Data":"61d2221d7874cbf20f6bb5f21c949b4b81d0303e48742f55c06d121b47ac5f31"} Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.761457 4715 scope.go:117] "RemoveContainer" containerID="7b89894e4cbe371d2f861bb14b093b8c1d5ab07f33cca597b2adc25ecabecf2d" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.761496 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vj7mq" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.799010 4715 scope.go:117] "RemoveContainer" containerID="6fdb12365d63bffca143f2580ad96b2cae63438d750c118301dee52d1a07d53d" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.823745 4715 scope.go:117] "RemoveContainer" containerID="c0f8ff110ecf8500b867b55c5ac32d8dcc4dba1982bf90340b67edf5bda7107d" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.858415 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4629f441-9cf2-469c-9815-5c1739cc429a" (UID: "4629f441-9cf2-469c-9815-5c1739cc429a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:25.946344 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4629f441-9cf2-469c-9815-5c1739cc429a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:26.089355 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vj7mq"] Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:26.092597 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vj7mq"] Dec 03 21:56:26 crc kubenswrapper[4715]: I1203 21:56:26.995614 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.164009 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-catalog-content\") pod \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.164134 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nx8l\" (UniqueName: \"kubernetes.io/projected/d0bce231-d2cd-44fe-b55e-ee830798ad3f-kube-api-access-8nx8l\") pod \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.164205 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-utilities\") pod \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\" (UID: \"d0bce231-d2cd-44fe-b55e-ee830798ad3f\") " Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.167086 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-utilities" (OuterVolumeSpecName: "utilities") pod "d0bce231-d2cd-44fe-b55e-ee830798ad3f" (UID: "d0bce231-d2cd-44fe-b55e-ee830798ad3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.173790 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0bce231-d2cd-44fe-b55e-ee830798ad3f-kube-api-access-8nx8l" (OuterVolumeSpecName: "kube-api-access-8nx8l") pod "d0bce231-d2cd-44fe-b55e-ee830798ad3f" (UID: "d0bce231-d2cd-44fe-b55e-ee830798ad3f"). InnerVolumeSpecName "kube-api-access-8nx8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.203431 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0bce231-d2cd-44fe-b55e-ee830798ad3f" (UID: "d0bce231-d2cd-44fe-b55e-ee830798ad3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.265442 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.265478 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nx8l\" (UniqueName: \"kubernetes.io/projected/d0bce231-d2cd-44fe-b55e-ee830798ad3f-kube-api-access-8nx8l\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.265489 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0bce231-d2cd-44fe-b55e-ee830798ad3f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.645125 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" path="/var/lib/kubelet/pods/4629f441-9cf2-469c-9815-5c1739cc429a/volumes" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.776069 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bx8qp" event={"ID":"d0bce231-d2cd-44fe-b55e-ee830798ad3f","Type":"ContainerDied","Data":"a771ffe9ef97739b13c343e66e55601d1eee0405dec854be1ce322a259d27a5d"} Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.776160 4715 scope.go:117] "RemoveContainer" containerID="a738cb6adab6591958c9421852fb59ef293e9355099621e40e106866f90c7b8c" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.776207 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bx8qp" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.795622 4715 scope.go:117] "RemoveContainer" containerID="659b9568b2c63ffc1e4a98199293e6c965fe03713f70cd1bfc67efcb91c6be42" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.808919 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx8qp"] Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.811722 4715 scope.go:117] "RemoveContainer" containerID="ccfbb72d50a27cc4a75ba0a1ee5bb2644fff3d31afac00d00137a464a56dfcb0" Dec 03 21:56:27 crc kubenswrapper[4715]: I1203 21:56:27.814171 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bx8qp"] Dec 03 21:56:29 crc kubenswrapper[4715]: I1203 21:56:29.644353 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" path="/var/lib/kubelet/pods/d0bce231-d2cd-44fe-b55e-ee830798ad3f/volumes" Dec 03 21:56:56 crc kubenswrapper[4715]: I1203 21:56:56.304234 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zfj7z"] Dec 03 21:56:56 crc kubenswrapper[4715]: I1203 21:56:56.307355 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" podUID="8b3ac689-17bc-489d-a467-1409ab8bd635" containerName="controller-manager" containerID="cri-o://f1a91605beb814f6f6eb16be07e1ef1302620d426e96338ef5c88504db0a0bca" gracePeriod=30 Dec 03 21:56:56 crc kubenswrapper[4715]: I1203 21:56:56.401106 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9"] Dec 03 21:56:56 crc kubenswrapper[4715]: I1203 21:56:56.401478 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" podUID="1ea0156d-4b30-4eb7-a285-a7becf5428ed" containerName="route-controller-manager" containerID="cri-o://37799e55bf1fed232c7710b7ced9fcd5e20a4d68e4b3467fe998d0364f0b898e" gracePeriod=30 Dec 03 21:56:56 crc kubenswrapper[4715]: I1203 21:56:56.808131 4715 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-zfj7z container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" start-of-body= Dec 03 21:56:56 crc kubenswrapper[4715]: I1203 21:56:56.808533 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" podUID="8b3ac689-17bc-489d-a467-1409ab8bd635" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" Dec 03 21:56:57 crc kubenswrapper[4715]: I1203 21:56:57.979634 4715 generic.go:334] "Generic (PLEG): container finished" podID="1ea0156d-4b30-4eb7-a285-a7becf5428ed" containerID="37799e55bf1fed232c7710b7ced9fcd5e20a4d68e4b3467fe998d0364f0b898e" exitCode=0 Dec 03 21:56:57 crc kubenswrapper[4715]: I1203 21:56:57.979888 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" event={"ID":"1ea0156d-4b30-4eb7-a285-a7becf5428ed","Type":"ContainerDied","Data":"37799e55bf1fed232c7710b7ced9fcd5e20a4d68e4b3467fe998d0364f0b898e"} Dec 03 21:56:57 crc kubenswrapper[4715]: I1203 21:56:57.981323 4715 generic.go:334] "Generic (PLEG): container finished" podID="8b3ac689-17bc-489d-a467-1409ab8bd635" containerID="f1a91605beb814f6f6eb16be07e1ef1302620d426e96338ef5c88504db0a0bca" exitCode=0 Dec 03 21:56:57 crc kubenswrapper[4715]: I1203 21:56:57.981357 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" event={"ID":"8b3ac689-17bc-489d-a467-1409ab8bd635","Type":"ContainerDied","Data":"f1a91605beb814f6f6eb16be07e1ef1302620d426e96338ef5c88504db0a0bca"} Dec 03 21:56:57 crc kubenswrapper[4715]: I1203 21:56:57.981382 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" event={"ID":"8b3ac689-17bc-489d-a467-1409ab8bd635","Type":"ContainerDied","Data":"ce96fa885c485d64191eae94b9546545c52708aaf5d8c53c7d6847b44dad289a"} Dec 03 21:56:57 crc kubenswrapper[4715]: I1203 21:56:57.981400 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce96fa885c485d64191eae94b9546545c52708aaf5d8c53c7d6847b44dad289a" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.037476 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.050742 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.080883 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-8647b84975-fvsj9"] Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081133 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerName="extract-content" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081148 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerName="extract-content" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081168 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b3ac689-17bc-489d-a467-1409ab8bd635" containerName="controller-manager" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081175 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b3ac689-17bc-489d-a467-1409ab8bd635" containerName="controller-manager" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081195 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea0156d-4b30-4eb7-a285-a7becf5428ed" containerName="route-controller-manager" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081203 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea0156d-4b30-4eb7-a285-a7becf5428ed" containerName="route-controller-manager" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081214 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerName="extract-utilities" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081220 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerName="extract-utilities" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081230 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerName="extract-content" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081238 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerName="extract-content" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081248 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerName="extract-utilities" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081255 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerName="extract-utilities" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081267 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" containerName="extract-utilities" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081274 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" containerName="extract-utilities" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081286 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" containerName="extract-content" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081292 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" containerName="extract-content" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081307 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081313 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081324 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081331 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081340 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081346 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081356 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081371 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081379 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerName="extract-content" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081385 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerName="extract-content" Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.081395 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerName="extract-utilities" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081401 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerName="extract-utilities" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081525 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eda92a8-2005-4801-befd-d1db15f1b2fa" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081538 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7912273-d7d2-4887-8525-e0c9e20d569a" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081547 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b3ac689-17bc-489d-a467-1409ab8bd635" containerName="controller-manager" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081562 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea0156d-4b30-4eb7-a285-a7becf5428ed" containerName="route-controller-manager" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081575 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0bce231-d2cd-44fe-b55e-ee830798ad3f" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.081582 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4629f441-9cf2-469c-9815-5c1739cc429a" containerName="registry-server" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.082006 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.092102 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-8647b84975-fvsj9"] Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.199738 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-8647b84975-fvsj9"] Dec 03 21:56:58 crc kubenswrapper[4715]: E1203 21:56:58.200324 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-x87zs proxy-ca-bundles serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" podUID="71538b6e-c16c-46bc-a3aa-878890c27f39" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.218552 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b3ac689-17bc-489d-a467-1409ab8bd635-serving-cert\") pod \"8b3ac689-17bc-489d-a467-1409ab8bd635\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.219984 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6mfb\" (UniqueName: \"kubernetes.io/projected/1ea0156d-4b30-4eb7-a285-a7becf5428ed-kube-api-access-v6mfb\") pod \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220019 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-config\") pod \"8b3ac689-17bc-489d-a467-1409ab8bd635\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220055 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ea0156d-4b30-4eb7-a285-a7becf5428ed-serving-cert\") pod \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220106 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-proxy-ca-bundles\") pod \"8b3ac689-17bc-489d-a467-1409ab8bd635\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220129 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-client-ca\") pod \"8b3ac689-17bc-489d-a467-1409ab8bd635\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220162 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-config\") pod \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220213 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sjw9\" (UniqueName: \"kubernetes.io/projected/8b3ac689-17bc-489d-a467-1409ab8bd635-kube-api-access-9sjw9\") pod \"8b3ac689-17bc-489d-a467-1409ab8bd635\" (UID: \"8b3ac689-17bc-489d-a467-1409ab8bd635\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220267 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-client-ca\") pod \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\" (UID: \"1ea0156d-4b30-4eb7-a285-a7becf5428ed\") " Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220488 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-proxy-ca-bundles\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220549 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x87zs\" (UniqueName: \"kubernetes.io/projected/71538b6e-c16c-46bc-a3aa-878890c27f39-kube-api-access-x87zs\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220576 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-client-ca\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220603 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-config\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220624 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71538b6e-c16c-46bc-a3aa-878890c27f39-serving-cert\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.220753 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "8b3ac689-17bc-489d-a467-1409ab8bd635" (UID: "8b3ac689-17bc-489d-a467-1409ab8bd635"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.221252 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-client-ca" (OuterVolumeSpecName: "client-ca") pod "8b3ac689-17bc-489d-a467-1409ab8bd635" (UID: "8b3ac689-17bc-489d-a467-1409ab8bd635"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.221577 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-config" (OuterVolumeSpecName: "config") pod "8b3ac689-17bc-489d-a467-1409ab8bd635" (UID: "8b3ac689-17bc-489d-a467-1409ab8bd635"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.221818 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-config" (OuterVolumeSpecName: "config") pod "1ea0156d-4b30-4eb7-a285-a7becf5428ed" (UID: "1ea0156d-4b30-4eb7-a285-a7becf5428ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.222560 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-client-ca" (OuterVolumeSpecName: "client-ca") pod "1ea0156d-4b30-4eb7-a285-a7becf5428ed" (UID: "1ea0156d-4b30-4eb7-a285-a7becf5428ed"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.225718 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b3ac689-17bc-489d-a467-1409ab8bd635-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8b3ac689-17bc-489d-a467-1409ab8bd635" (UID: "8b3ac689-17bc-489d-a467-1409ab8bd635"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.225791 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b3ac689-17bc-489d-a467-1409ab8bd635-kube-api-access-9sjw9" (OuterVolumeSpecName: "kube-api-access-9sjw9") pod "8b3ac689-17bc-489d-a467-1409ab8bd635" (UID: "8b3ac689-17bc-489d-a467-1409ab8bd635"). InnerVolumeSpecName "kube-api-access-9sjw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.229354 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ea0156d-4b30-4eb7-a285-a7becf5428ed-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1ea0156d-4b30-4eb7-a285-a7becf5428ed" (UID: "1ea0156d-4b30-4eb7-a285-a7becf5428ed"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.239436 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea0156d-4b30-4eb7-a285-a7becf5428ed-kube-api-access-v6mfb" (OuterVolumeSpecName: "kube-api-access-v6mfb") pod "1ea0156d-4b30-4eb7-a285-a7becf5428ed" (UID: "1ea0156d-4b30-4eb7-a285-a7becf5428ed"). InnerVolumeSpecName "kube-api-access-v6mfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322315 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-config\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322410 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71538b6e-c16c-46bc-a3aa-878890c27f39-serving-cert\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322478 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-proxy-ca-bundles\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322531 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x87zs\" (UniqueName: \"kubernetes.io/projected/71538b6e-c16c-46bc-a3aa-878890c27f39-kube-api-access-x87zs\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322565 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-client-ca\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322638 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sjw9\" (UniqueName: \"kubernetes.io/projected/8b3ac689-17bc-489d-a467-1409ab8bd635-kube-api-access-9sjw9\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322655 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322667 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b3ac689-17bc-489d-a467-1409ab8bd635-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322678 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6mfb\" (UniqueName: \"kubernetes.io/projected/1ea0156d-4b30-4eb7-a285-a7becf5428ed-kube-api-access-v6mfb\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322689 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322702 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1ea0156d-4b30-4eb7-a285-a7becf5428ed-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322713 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322722 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b3ac689-17bc-489d-a467-1409ab8bd635-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.322731 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea0156d-4b30-4eb7-a285-a7becf5428ed-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.323761 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-client-ca\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.324651 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-config\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.326328 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-proxy-ca-bundles\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.328929 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71538b6e-c16c-46bc-a3aa-878890c27f39-serving-cert\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.343968 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x87zs\" (UniqueName: \"kubernetes.io/projected/71538b6e-c16c-46bc-a3aa-878890c27f39-kube-api-access-x87zs\") pod \"controller-manager-8647b84975-fvsj9\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.990883 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" event={"ID":"1ea0156d-4b30-4eb7-a285-a7becf5428ed","Type":"ContainerDied","Data":"b749502ea9bd91bf7f40761f5198c20a7dedfb03684d9c95e899e41ecc74911b"} Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.990935 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-zfj7z" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.990973 4715 scope.go:117] "RemoveContainer" containerID="37799e55bf1fed232c7710b7ced9fcd5e20a4d68e4b3467fe998d0364f0b898e" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.990990 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:58 crc kubenswrapper[4715]: I1203 21:56:58.991120 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.005815 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.038272 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zfj7z"] Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.049622 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-zfj7z"] Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.054235 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9"] Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.058232 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-48bw9"] Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.136716 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71538b6e-c16c-46bc-a3aa-878890c27f39-serving-cert\") pod \"71538b6e-c16c-46bc-a3aa-878890c27f39\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.136788 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-client-ca\") pod \"71538b6e-c16c-46bc-a3aa-878890c27f39\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.136830 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-proxy-ca-bundles\") pod \"71538b6e-c16c-46bc-a3aa-878890c27f39\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.136982 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x87zs\" (UniqueName: \"kubernetes.io/projected/71538b6e-c16c-46bc-a3aa-878890c27f39-kube-api-access-x87zs\") pod \"71538b6e-c16c-46bc-a3aa-878890c27f39\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.137024 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-config\") pod \"71538b6e-c16c-46bc-a3aa-878890c27f39\" (UID: \"71538b6e-c16c-46bc-a3aa-878890c27f39\") " Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.139265 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-config" (OuterVolumeSpecName: "config") pod "71538b6e-c16c-46bc-a3aa-878890c27f39" (UID: "71538b6e-c16c-46bc-a3aa-878890c27f39"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.139974 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-client-ca" (OuterVolumeSpecName: "client-ca") pod "71538b6e-c16c-46bc-a3aa-878890c27f39" (UID: "71538b6e-c16c-46bc-a3aa-878890c27f39"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.140277 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "71538b6e-c16c-46bc-a3aa-878890c27f39" (UID: "71538b6e-c16c-46bc-a3aa-878890c27f39"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.151900 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71538b6e-c16c-46bc-a3aa-878890c27f39-kube-api-access-x87zs" (OuterVolumeSpecName: "kube-api-access-x87zs") pod "71538b6e-c16c-46bc-a3aa-878890c27f39" (UID: "71538b6e-c16c-46bc-a3aa-878890c27f39"). InnerVolumeSpecName "kube-api-access-x87zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.156690 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71538b6e-c16c-46bc-a3aa-878890c27f39-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "71538b6e-c16c-46bc-a3aa-878890c27f39" (UID: "71538b6e-c16c-46bc-a3aa-878890c27f39"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.238883 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x87zs\" (UniqueName: \"kubernetes.io/projected/71538b6e-c16c-46bc-a3aa-878890c27f39-kube-api-access-x87zs\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.238933 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-config\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.238943 4715 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71538b6e-c16c-46bc-a3aa-878890c27f39-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.238952 4715 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.238961 4715 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/71538b6e-c16c-46bc-a3aa-878890c27f39-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.651079 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea0156d-4b30-4eb7-a285-a7becf5428ed" path="/var/lib/kubelet/pods/1ea0156d-4b30-4eb7-a285-a7becf5428ed/volumes" Dec 03 21:56:59 crc kubenswrapper[4715]: I1203 21:56:59.670467 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b3ac689-17bc-489d-a467-1409ab8bd635" path="/var/lib/kubelet/pods/8b3ac689-17bc-489d-a467-1409ab8bd635/volumes" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.000117 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-8647b84975-fvsj9" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.050872 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq"] Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.052611 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.056567 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.056744 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.056905 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.057408 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-8647b84975-fvsj9"] Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.061888 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-8647b84975-fvsj9"] Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.068007 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.068045 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.068329 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.077538 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq"] Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.083080 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.153931 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-client-ca\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.154049 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-proxy-ca-bundles\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.154102 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-serving-cert\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.154139 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-config\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.154193 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7pqj\" (UniqueName: \"kubernetes.io/projected/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-kube-api-access-r7pqj\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.255732 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-client-ca\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.255812 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-proxy-ca-bundles\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.255838 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-serving-cert\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.255855 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-config\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.255883 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7pqj\" (UniqueName: \"kubernetes.io/projected/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-kube-api-access-r7pqj\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.257046 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-client-ca\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.257386 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-config\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.259157 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-proxy-ca-bundles\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.269641 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-serving-cert\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.278481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7pqj\" (UniqueName: \"kubernetes.io/projected/448b3ea8-5f49-4aa0-b385-6c6bd8aeb206-kube-api-access-r7pqj\") pod \"controller-manager-7d9c6dc78b-2q5fq\" (UID: \"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206\") " pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.372814 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.678774 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq"] Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.867155 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b"] Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.868440 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.875906 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.878771 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.879033 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.879343 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.879571 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b"] Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.879605 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.879711 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.966491 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxxm9\" (UniqueName: \"kubernetes.io/projected/9f26b963-8f88-4300-b1c2-e7441c5bcf64-kube-api-access-xxxm9\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.966615 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f26b963-8f88-4300-b1c2-e7441c5bcf64-serving-cert\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.966874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f26b963-8f88-4300-b1c2-e7441c5bcf64-client-ca\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:00 crc kubenswrapper[4715]: I1203 21:57:00.967062 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f26b963-8f88-4300-b1c2-e7441c5bcf64-config\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.009326 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" event={"ID":"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206","Type":"ContainerStarted","Data":"41bb53ad7b4445a1e12852c15abfad000ad746297702aa09e86cc88eb827ee88"} Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.068821 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxxm9\" (UniqueName: \"kubernetes.io/projected/9f26b963-8f88-4300-b1c2-e7441c5bcf64-kube-api-access-xxxm9\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.068896 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f26b963-8f88-4300-b1c2-e7441c5bcf64-serving-cert\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.068979 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f26b963-8f88-4300-b1c2-e7441c5bcf64-client-ca\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.069053 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f26b963-8f88-4300-b1c2-e7441c5bcf64-config\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.071642 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9f26b963-8f88-4300-b1c2-e7441c5bcf64-client-ca\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.072807 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f26b963-8f88-4300-b1c2-e7441c5bcf64-config\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.083859 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9f26b963-8f88-4300-b1c2-e7441c5bcf64-serving-cert\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.102026 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxxm9\" (UniqueName: \"kubernetes.io/projected/9f26b963-8f88-4300-b1c2-e7441c5bcf64-kube-api-access-xxxm9\") pod \"route-controller-manager-568b76d76b-nkz8b\" (UID: \"9f26b963-8f88-4300-b1c2-e7441c5bcf64\") " pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:01 crc kubenswrapper[4715]: I1203 21:57:01.204387 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:02 crc kubenswrapper[4715]: I1203 21:57:01.645893 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71538b6e-c16c-46bc-a3aa-878890c27f39" path="/var/lib/kubelet/pods/71538b6e-c16c-46bc-a3aa-878890c27f39/volumes" Dec 03 21:57:02 crc kubenswrapper[4715]: I1203 21:57:02.020220 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" event={"ID":"448b3ea8-5f49-4aa0-b385-6c6bd8aeb206","Type":"ContainerStarted","Data":"bc5457a63fcb0a7c7ba3ff7856964877ab9266a39a450319f2ee246f999e3714"} Dec 03 21:57:02 crc kubenswrapper[4715]: I1203 21:57:02.020654 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:02 crc kubenswrapper[4715]: I1203 21:57:02.030442 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" Dec 03 21:57:02 crc kubenswrapper[4715]: I1203 21:57:02.053530 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d9c6dc78b-2q5fq" podStartSLOduration=4.053469387 podStartE2EDuration="4.053469387s" podCreationTimestamp="2025-12-03 21:56:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:57:02.046699444 +0000 UTC m=+918.789410079" watchObservedRunningTime="2025-12-03 21:57:02.053469387 +0000 UTC m=+918.796180012" Dec 03 21:57:03 crc kubenswrapper[4715]: I1203 21:57:03.136667 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b"] Dec 03 21:57:04 crc kubenswrapper[4715]: I1203 21:57:04.033597 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" event={"ID":"9f26b963-8f88-4300-b1c2-e7441c5bcf64","Type":"ContainerStarted","Data":"f8a9467e354a0668957cdc50362b7b38f9543f85a7b538f55ce6df571a2f0d14"} Dec 03 21:57:04 crc kubenswrapper[4715]: I1203 21:57:04.033668 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" event={"ID":"9f26b963-8f88-4300-b1c2-e7441c5bcf64","Type":"ContainerStarted","Data":"5c59acb876594ebf71f1944230a13a919be4423fb4e96a9726b9d05bbc07f3f6"} Dec 03 21:57:04 crc kubenswrapper[4715]: I1203 21:57:04.058920 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" podStartSLOduration=8.058898143 podStartE2EDuration="8.058898143s" podCreationTimestamp="2025-12-03 21:56:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 21:57:04.056351294 +0000 UTC m=+920.799061919" watchObservedRunningTime="2025-12-03 21:57:04.058898143 +0000 UTC m=+920.801608728" Dec 03 21:57:05 crc kubenswrapper[4715]: I1203 21:57:05.038362 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:05 crc kubenswrapper[4715]: I1203 21:57:05.044659 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-568b76d76b-nkz8b" Dec 03 21:57:35 crc kubenswrapper[4715]: I1203 21:57:35.160801 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:57:35 crc kubenswrapper[4715]: I1203 21:57:35.161489 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:57:44 crc kubenswrapper[4715]: I1203 21:57:44.042644 4715 scope.go:117] "RemoveContainer" containerID="f1a91605beb814f6f6eb16be07e1ef1302620d426e96338ef5c88504db0a0bca" Dec 03 21:58:05 crc kubenswrapper[4715]: I1203 21:58:05.159952 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:58:05 crc kubenswrapper[4715]: I1203 21:58:05.162344 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.159924 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.160669 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.160734 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.161554 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e257c66c2cc6aa263755e73d7447050e7965b799fc91ed32403621c7f701856d"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.161666 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://e257c66c2cc6aa263755e73d7447050e7965b799fc91ed32403621c7f701856d" gracePeriod=600 Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.694320 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="e257c66c2cc6aa263755e73d7447050e7965b799fc91ed32403621c7f701856d" exitCode=0 Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.694386 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"e257c66c2cc6aa263755e73d7447050e7965b799fc91ed32403621c7f701856d"} Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.694833 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"090018a715ee87b87ef301bea6ea6579973e0fe07aad087b1f463ba8cdf860ce"} Dec 03 21:58:35 crc kubenswrapper[4715]: I1203 21:58:35.694866 4715 scope.go:117] "RemoveContainer" containerID="40f9d0ad19f2130ff948a7243a29e414da6e6f2dd3f614df95d73b2addc6413d" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.198557 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc"] Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.199917 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.202625 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.203422 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.213840 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc"] Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.273828 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f97b6863-0677-43be-9fcd-d6c18bed4f90-secret-volume\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.273900 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tkns\" (UniqueName: \"kubernetes.io/projected/f97b6863-0677-43be-9fcd-d6c18bed4f90-kube-api-access-5tkns\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.274013 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f97b6863-0677-43be-9fcd-d6c18bed4f90-config-volume\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.375026 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f97b6863-0677-43be-9fcd-d6c18bed4f90-secret-volume\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.375096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tkns\" (UniqueName: \"kubernetes.io/projected/f97b6863-0677-43be-9fcd-d6c18bed4f90-kube-api-access-5tkns\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.375164 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f97b6863-0677-43be-9fcd-d6c18bed4f90-config-volume\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.376326 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f97b6863-0677-43be-9fcd-d6c18bed4f90-config-volume\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.386402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f97b6863-0677-43be-9fcd-d6c18bed4f90-secret-volume\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.397690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tkns\" (UniqueName: \"kubernetes.io/projected/f97b6863-0677-43be-9fcd-d6c18bed4f90-kube-api-access-5tkns\") pod \"collect-profiles-29413320-49jkc\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.537528 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:00 crc kubenswrapper[4715]: I1203 22:00:00.825696 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc"] Dec 03 22:00:01 crc kubenswrapper[4715]: I1203 22:00:01.302710 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" event={"ID":"f97b6863-0677-43be-9fcd-d6c18bed4f90","Type":"ContainerStarted","Data":"31af019dfdcbee4f13dee30c1e68c7743db635344c3ded4e2cfc9579aa62b8f2"} Dec 03 22:00:02 crc kubenswrapper[4715]: I1203 22:00:02.311951 4715 generic.go:334] "Generic (PLEG): container finished" podID="f97b6863-0677-43be-9fcd-d6c18bed4f90" containerID="c2af3a574582e17f741374ded3450b91b586e46fa7b6cdd65812b502889e1a68" exitCode=0 Dec 03 22:00:02 crc kubenswrapper[4715]: I1203 22:00:02.312051 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" event={"ID":"f97b6863-0677-43be-9fcd-d6c18bed4f90","Type":"ContainerDied","Data":"c2af3a574582e17f741374ded3450b91b586e46fa7b6cdd65812b502889e1a68"} Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.598600 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.732250 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tkns\" (UniqueName: \"kubernetes.io/projected/f97b6863-0677-43be-9fcd-d6c18bed4f90-kube-api-access-5tkns\") pod \"f97b6863-0677-43be-9fcd-d6c18bed4f90\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.732324 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f97b6863-0677-43be-9fcd-d6c18bed4f90-secret-volume\") pod \"f97b6863-0677-43be-9fcd-d6c18bed4f90\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.732434 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f97b6863-0677-43be-9fcd-d6c18bed4f90-config-volume\") pod \"f97b6863-0677-43be-9fcd-d6c18bed4f90\" (UID: \"f97b6863-0677-43be-9fcd-d6c18bed4f90\") " Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.733763 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f97b6863-0677-43be-9fcd-d6c18bed4f90-config-volume" (OuterVolumeSpecName: "config-volume") pod "f97b6863-0677-43be-9fcd-d6c18bed4f90" (UID: "f97b6863-0677-43be-9fcd-d6c18bed4f90"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.739039 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f97b6863-0677-43be-9fcd-d6c18bed4f90-kube-api-access-5tkns" (OuterVolumeSpecName: "kube-api-access-5tkns") pod "f97b6863-0677-43be-9fcd-d6c18bed4f90" (UID: "f97b6863-0677-43be-9fcd-d6c18bed4f90"). InnerVolumeSpecName "kube-api-access-5tkns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.740190 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f97b6863-0677-43be-9fcd-d6c18bed4f90-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f97b6863-0677-43be-9fcd-d6c18bed4f90" (UID: "f97b6863-0677-43be-9fcd-d6c18bed4f90"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.834556 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f97b6863-0677-43be-9fcd-d6c18bed4f90-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.834612 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f97b6863-0677-43be-9fcd-d6c18bed4f90-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 22:00:03 crc kubenswrapper[4715]: I1203 22:00:03.834626 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tkns\" (UniqueName: \"kubernetes.io/projected/f97b6863-0677-43be-9fcd-d6c18bed4f90-kube-api-access-5tkns\") on node \"crc\" DevicePath \"\"" Dec 03 22:00:04 crc kubenswrapper[4715]: I1203 22:00:04.327074 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" event={"ID":"f97b6863-0677-43be-9fcd-d6c18bed4f90","Type":"ContainerDied","Data":"31af019dfdcbee4f13dee30c1e68c7743db635344c3ded4e2cfc9579aa62b8f2"} Dec 03 22:00:04 crc kubenswrapper[4715]: I1203 22:00:04.327112 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31af019dfdcbee4f13dee30c1e68c7743db635344c3ded4e2cfc9579aa62b8f2" Dec 03 22:00:04 crc kubenswrapper[4715]: I1203 22:00:04.327224 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc" Dec 03 22:00:35 crc kubenswrapper[4715]: I1203 22:00:35.160193 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:00:35 crc kubenswrapper[4715]: I1203 22:00:35.161010 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:01:05 crc kubenswrapper[4715]: I1203 22:01:05.160036 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:01:05 crc kubenswrapper[4715]: I1203 22:01:05.160737 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:01:35 crc kubenswrapper[4715]: I1203 22:01:35.159784 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:01:35 crc kubenswrapper[4715]: I1203 22:01:35.160451 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:01:35 crc kubenswrapper[4715]: I1203 22:01:35.160576 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:01:35 crc kubenswrapper[4715]: I1203 22:01:35.161388 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"090018a715ee87b87ef301bea6ea6579973e0fe07aad087b1f463ba8cdf860ce"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:01:35 crc kubenswrapper[4715]: I1203 22:01:35.161497 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://090018a715ee87b87ef301bea6ea6579973e0fe07aad087b1f463ba8cdf860ce" gracePeriod=600 Dec 03 22:01:36 crc kubenswrapper[4715]: I1203 22:01:36.007222 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="090018a715ee87b87ef301bea6ea6579973e0fe07aad087b1f463ba8cdf860ce" exitCode=0 Dec 03 22:01:36 crc kubenswrapper[4715]: I1203 22:01:36.007316 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"090018a715ee87b87ef301bea6ea6579973e0fe07aad087b1f463ba8cdf860ce"} Dec 03 22:01:36 crc kubenswrapper[4715]: I1203 22:01:36.007938 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"f5f277a5e5960cbb4f3e306224a9016bab30508a2102823dbd0cc14c7e19cc97"} Dec 03 22:01:36 crc kubenswrapper[4715]: I1203 22:01:36.007974 4715 scope.go:117] "RemoveContainer" containerID="e257c66c2cc6aa263755e73d7447050e7965b799fc91ed32403621c7f701856d" Dec 03 22:03:35 crc kubenswrapper[4715]: I1203 22:03:35.161355 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:03:35 crc kubenswrapper[4715]: I1203 22:03:35.162393 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:04:05 crc kubenswrapper[4715]: I1203 22:04:05.160391 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:04:05 crc kubenswrapper[4715]: I1203 22:04:05.161062 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:04:35 crc kubenswrapper[4715]: I1203 22:04:35.160434 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:04:35 crc kubenswrapper[4715]: I1203 22:04:35.161298 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:04:35 crc kubenswrapper[4715]: I1203 22:04:35.161368 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:04:35 crc kubenswrapper[4715]: I1203 22:04:35.162291 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f5f277a5e5960cbb4f3e306224a9016bab30508a2102823dbd0cc14c7e19cc97"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:04:35 crc kubenswrapper[4715]: I1203 22:04:35.162387 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://f5f277a5e5960cbb4f3e306224a9016bab30508a2102823dbd0cc14c7e19cc97" gracePeriod=600 Dec 03 22:04:35 crc kubenswrapper[4715]: I1203 22:04:35.387967 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="f5f277a5e5960cbb4f3e306224a9016bab30508a2102823dbd0cc14c7e19cc97" exitCode=0 Dec 03 22:04:35 crc kubenswrapper[4715]: I1203 22:04:35.388021 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"f5f277a5e5960cbb4f3e306224a9016bab30508a2102823dbd0cc14c7e19cc97"} Dec 03 22:04:35 crc kubenswrapper[4715]: I1203 22:04:35.388062 4715 scope.go:117] "RemoveContainer" containerID="090018a715ee87b87ef301bea6ea6579973e0fe07aad087b1f463ba8cdf860ce" Dec 03 22:04:36 crc kubenswrapper[4715]: I1203 22:04:36.397557 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02"} Dec 03 22:06:28 crc kubenswrapper[4715]: I1203 22:06:28.896388 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8lkk5"] Dec 03 22:06:28 crc kubenswrapper[4715]: E1203 22:06:28.897651 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97b6863-0677-43be-9fcd-d6c18bed4f90" containerName="collect-profiles" Dec 03 22:06:28 crc kubenswrapper[4715]: I1203 22:06:28.897685 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97b6863-0677-43be-9fcd-d6c18bed4f90" containerName="collect-profiles" Dec 03 22:06:28 crc kubenswrapper[4715]: I1203 22:06:28.897923 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97b6863-0677-43be-9fcd-d6c18bed4f90" containerName="collect-profiles" Dec 03 22:06:28 crc kubenswrapper[4715]: I1203 22:06:28.899653 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:28 crc kubenswrapper[4715]: I1203 22:06:28.910820 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8lkk5"] Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.049371 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtvn9\" (UniqueName: \"kubernetes.io/projected/4e9933f6-866c-44d4-b902-d37d4327e9d9-kube-api-access-xtvn9\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.049469 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-catalog-content\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.049710 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-utilities\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.150640 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-utilities\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.150805 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtvn9\" (UniqueName: \"kubernetes.io/projected/4e9933f6-866c-44d4-b902-d37d4327e9d9-kube-api-access-xtvn9\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.150853 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-catalog-content\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.151761 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-utilities\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.151776 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-catalog-content\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.193537 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtvn9\" (UniqueName: \"kubernetes.io/projected/4e9933f6-866c-44d4-b902-d37d4327e9d9-kube-api-access-xtvn9\") pod \"certified-operators-8lkk5\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.229585 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:29 crc kubenswrapper[4715]: I1203 22:06:29.527292 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8lkk5"] Dec 03 22:06:30 crc kubenswrapper[4715]: I1203 22:06:30.191618 4715 generic.go:334] "Generic (PLEG): container finished" podID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerID="018a06bbbd9305b4dd4d12ca5073c7eb9c57f97a83fec3a36f9e186873dbe491" exitCode=0 Dec 03 22:06:30 crc kubenswrapper[4715]: I1203 22:06:30.191757 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lkk5" event={"ID":"4e9933f6-866c-44d4-b902-d37d4327e9d9","Type":"ContainerDied","Data":"018a06bbbd9305b4dd4d12ca5073c7eb9c57f97a83fec3a36f9e186873dbe491"} Dec 03 22:06:30 crc kubenswrapper[4715]: I1203 22:06:30.192170 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lkk5" event={"ID":"4e9933f6-866c-44d4-b902-d37d4327e9d9","Type":"ContainerStarted","Data":"9e097fe3900a0724b6d54402ad43b1a773f2ddb6f9e3bbad8b3cfc40e3974928"} Dec 03 22:06:30 crc kubenswrapper[4715]: I1203 22:06:30.194242 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:06:31 crc kubenswrapper[4715]: I1203 22:06:31.200646 4715 generic.go:334] "Generic (PLEG): container finished" podID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerID="68099e88c97ded7960e64ae5dadd1e7b18def0ab18577e553df12c8e38667310" exitCode=0 Dec 03 22:06:31 crc kubenswrapper[4715]: I1203 22:06:31.200761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lkk5" event={"ID":"4e9933f6-866c-44d4-b902-d37d4327e9d9","Type":"ContainerDied","Data":"68099e88c97ded7960e64ae5dadd1e7b18def0ab18577e553df12c8e38667310"} Dec 03 22:06:32 crc kubenswrapper[4715]: I1203 22:06:32.217936 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lkk5" event={"ID":"4e9933f6-866c-44d4-b902-d37d4327e9d9","Type":"ContainerStarted","Data":"6b41dce2d532edd092a98d8cbe2f56ea91d9ab4b40049cb89246e03e0b7f7022"} Dec 03 22:06:32 crc kubenswrapper[4715]: I1203 22:06:32.245462 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8lkk5" podStartSLOduration=2.8360203520000002 podStartE2EDuration="4.245435864s" podCreationTimestamp="2025-12-03 22:06:28 +0000 UTC" firstStartedPulling="2025-12-03 22:06:30.193621972 +0000 UTC m=+1486.936332607" lastFinishedPulling="2025-12-03 22:06:31.603037494 +0000 UTC m=+1488.345748119" observedRunningTime="2025-12-03 22:06:32.238271893 +0000 UTC m=+1488.980982528" watchObservedRunningTime="2025-12-03 22:06:32.245435864 +0000 UTC m=+1488.988146489" Dec 03 22:06:35 crc kubenswrapper[4715]: I1203 22:06:35.160679 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:06:35 crc kubenswrapper[4715]: I1203 22:06:35.161254 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:06:39 crc kubenswrapper[4715]: I1203 22:06:39.230359 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:39 crc kubenswrapper[4715]: I1203 22:06:39.230801 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:39 crc kubenswrapper[4715]: I1203 22:06:39.295281 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:39 crc kubenswrapper[4715]: I1203 22:06:39.344526 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:39 crc kubenswrapper[4715]: I1203 22:06:39.533878 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8lkk5"] Dec 03 22:06:41 crc kubenswrapper[4715]: I1203 22:06:41.277090 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8lkk5" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerName="registry-server" containerID="cri-o://6b41dce2d532edd092a98d8cbe2f56ea91d9ab4b40049cb89246e03e0b7f7022" gracePeriod=2 Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.290109 4715 generic.go:334] "Generic (PLEG): container finished" podID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerID="6b41dce2d532edd092a98d8cbe2f56ea91d9ab4b40049cb89246e03e0b7f7022" exitCode=0 Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.290190 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lkk5" event={"ID":"4e9933f6-866c-44d4-b902-d37d4327e9d9","Type":"ContainerDied","Data":"6b41dce2d532edd092a98d8cbe2f56ea91d9ab4b40049cb89246e03e0b7f7022"} Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.444348 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.585736 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-catalog-content\") pod \"4e9933f6-866c-44d4-b902-d37d4327e9d9\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.585813 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtvn9\" (UniqueName: \"kubernetes.io/projected/4e9933f6-866c-44d4-b902-d37d4327e9d9-kube-api-access-xtvn9\") pod \"4e9933f6-866c-44d4-b902-d37d4327e9d9\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.585871 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-utilities\") pod \"4e9933f6-866c-44d4-b902-d37d4327e9d9\" (UID: \"4e9933f6-866c-44d4-b902-d37d4327e9d9\") " Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.586845 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-utilities" (OuterVolumeSpecName: "utilities") pod "4e9933f6-866c-44d4-b902-d37d4327e9d9" (UID: "4e9933f6-866c-44d4-b902-d37d4327e9d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.594748 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e9933f6-866c-44d4-b902-d37d4327e9d9-kube-api-access-xtvn9" (OuterVolumeSpecName: "kube-api-access-xtvn9") pod "4e9933f6-866c-44d4-b902-d37d4327e9d9" (UID: "4e9933f6-866c-44d4-b902-d37d4327e9d9"). InnerVolumeSpecName "kube-api-access-xtvn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.632318 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4e9933f6-866c-44d4-b902-d37d4327e9d9" (UID: "4e9933f6-866c-44d4-b902-d37d4327e9d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.687035 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.687088 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtvn9\" (UniqueName: \"kubernetes.io/projected/4e9933f6-866c-44d4-b902-d37d4327e9d9-kube-api-access-xtvn9\") on node \"crc\" DevicePath \"\"" Dec 03 22:06:43 crc kubenswrapper[4715]: I1203 22:06:43.687100 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4e9933f6-866c-44d4-b902-d37d4327e9d9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:06:44 crc kubenswrapper[4715]: I1203 22:06:44.301923 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8lkk5" event={"ID":"4e9933f6-866c-44d4-b902-d37d4327e9d9","Type":"ContainerDied","Data":"9e097fe3900a0724b6d54402ad43b1a773f2ddb6f9e3bbad8b3cfc40e3974928"} Dec 03 22:06:44 crc kubenswrapper[4715]: I1203 22:06:44.302024 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8lkk5" Dec 03 22:06:44 crc kubenswrapper[4715]: I1203 22:06:44.302346 4715 scope.go:117] "RemoveContainer" containerID="6b41dce2d532edd092a98d8cbe2f56ea91d9ab4b40049cb89246e03e0b7f7022" Dec 03 22:06:44 crc kubenswrapper[4715]: I1203 22:06:44.335146 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8lkk5"] Dec 03 22:06:44 crc kubenswrapper[4715]: I1203 22:06:44.339672 4715 scope.go:117] "RemoveContainer" containerID="68099e88c97ded7960e64ae5dadd1e7b18def0ab18577e553df12c8e38667310" Dec 03 22:06:44 crc kubenswrapper[4715]: I1203 22:06:44.343390 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8lkk5"] Dec 03 22:06:44 crc kubenswrapper[4715]: I1203 22:06:44.379607 4715 scope.go:117] "RemoveContainer" containerID="018a06bbbd9305b4dd4d12ca5073c7eb9c57f97a83fec3a36f9e186873dbe491" Dec 03 22:06:45 crc kubenswrapper[4715]: I1203 22:06:45.645910 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" path="/var/lib/kubelet/pods/4e9933f6-866c-44d4-b902-d37d4327e9d9/volumes" Dec 03 22:07:05 crc kubenswrapper[4715]: I1203 22:07:05.159783 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:07:05 crc kubenswrapper[4715]: I1203 22:07:05.160483 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.159896 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.160551 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.160624 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.161503 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.161624 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" gracePeriod=600 Dec 03 22:07:35 crc kubenswrapper[4715]: E1203 22:07:35.283876 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.631728 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" exitCode=0 Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.631800 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02"} Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.631899 4715 scope.go:117] "RemoveContainer" containerID="f5f277a5e5960cbb4f3e306224a9016bab30508a2102823dbd0cc14c7e19cc97" Dec 03 22:07:35 crc kubenswrapper[4715]: I1203 22:07:35.632979 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:07:35 crc kubenswrapper[4715]: E1203 22:07:35.633592 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:07:49 crc kubenswrapper[4715]: I1203 22:07:49.634906 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:07:49 crc kubenswrapper[4715]: E1203 22:07:49.635697 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:08:03 crc kubenswrapper[4715]: I1203 22:08:03.638384 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:08:03 crc kubenswrapper[4715]: E1203 22:08:03.639104 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:08:15 crc kubenswrapper[4715]: I1203 22:08:15.634218 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:08:15 crc kubenswrapper[4715]: E1203 22:08:15.635046 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.605540 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rnccs"] Dec 03 22:08:21 crc kubenswrapper[4715]: E1203 22:08:21.605962 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerName="registry-server" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.605974 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerName="registry-server" Dec 03 22:08:21 crc kubenswrapper[4715]: E1203 22:08:21.605988 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerName="extract-content" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.605993 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerName="extract-content" Dec 03 22:08:21 crc kubenswrapper[4715]: E1203 22:08:21.606009 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerName="extract-utilities" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.606015 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerName="extract-utilities" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.606103 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e9933f6-866c-44d4-b902-d37d4327e9d9" containerName="registry-server" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.606459 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rnccs" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.619495 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.619835 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.619868 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-8schp" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.623181 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rnccs"] Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.630884 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-4bvpr"] Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.631842 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-4bvpr" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.636561 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-68trn" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.659585 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-4bvpr"] Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.663522 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dgskj"] Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.665471 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.667702 4715 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-sjhxc" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.672155 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dgskj"] Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.738531 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwch7\" (UniqueName: \"kubernetes.io/projected/ce5706ce-67d9-4e61-8051-03113e8c3ac4-kube-api-access-rwch7\") pod \"cert-manager-5b446d88c5-4bvpr\" (UID: \"ce5706ce-67d9-4e61-8051-03113e8c3ac4\") " pod="cert-manager/cert-manager-5b446d88c5-4bvpr" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.738606 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29mz2\" (UniqueName: \"kubernetes.io/projected/bec62745-cf3a-4bc6-9fc6-34c8a56301ce-kube-api-access-29mz2\") pod \"cert-manager-cainjector-7f985d654d-rnccs\" (UID: \"bec62745-cf3a-4bc6-9fc6-34c8a56301ce\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rnccs" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.839635 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwddx\" (UniqueName: \"kubernetes.io/projected/30ac245b-49dd-48b9-878e-e06008f921b7-kube-api-access-lwddx\") pod \"cert-manager-webhook-5655c58dd6-dgskj\" (UID: \"30ac245b-49dd-48b9-878e-e06008f921b7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.839737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwch7\" (UniqueName: \"kubernetes.io/projected/ce5706ce-67d9-4e61-8051-03113e8c3ac4-kube-api-access-rwch7\") pod \"cert-manager-5b446d88c5-4bvpr\" (UID: \"ce5706ce-67d9-4e61-8051-03113e8c3ac4\") " pod="cert-manager/cert-manager-5b446d88c5-4bvpr" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.839771 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29mz2\" (UniqueName: \"kubernetes.io/projected/bec62745-cf3a-4bc6-9fc6-34c8a56301ce-kube-api-access-29mz2\") pod \"cert-manager-cainjector-7f985d654d-rnccs\" (UID: \"bec62745-cf3a-4bc6-9fc6-34c8a56301ce\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rnccs" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.873773 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29mz2\" (UniqueName: \"kubernetes.io/projected/bec62745-cf3a-4bc6-9fc6-34c8a56301ce-kube-api-access-29mz2\") pod \"cert-manager-cainjector-7f985d654d-rnccs\" (UID: \"bec62745-cf3a-4bc6-9fc6-34c8a56301ce\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-rnccs" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.875405 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwch7\" (UniqueName: \"kubernetes.io/projected/ce5706ce-67d9-4e61-8051-03113e8c3ac4-kube-api-access-rwch7\") pod \"cert-manager-5b446d88c5-4bvpr\" (UID: \"ce5706ce-67d9-4e61-8051-03113e8c3ac4\") " pod="cert-manager/cert-manager-5b446d88c5-4bvpr" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.923525 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-rnccs" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.940835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwddx\" (UniqueName: \"kubernetes.io/projected/30ac245b-49dd-48b9-878e-e06008f921b7-kube-api-access-lwddx\") pod \"cert-manager-webhook-5655c58dd6-dgskj\" (UID: \"30ac245b-49dd-48b9-878e-e06008f921b7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.955556 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-4bvpr" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.961786 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwddx\" (UniqueName: \"kubernetes.io/projected/30ac245b-49dd-48b9-878e-e06008f921b7-kube-api-access-lwddx\") pod \"cert-manager-webhook-5655c58dd6-dgskj\" (UID: \"30ac245b-49dd-48b9-878e-e06008f921b7\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" Dec 03 22:08:21 crc kubenswrapper[4715]: I1203 22:08:21.980748 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" Dec 03 22:08:22 crc kubenswrapper[4715]: I1203 22:08:22.210251 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-4bvpr"] Dec 03 22:08:22 crc kubenswrapper[4715]: I1203 22:08:22.283866 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-dgskj"] Dec 03 22:08:22 crc kubenswrapper[4715]: W1203 22:08:22.284101 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30ac245b_49dd_48b9_878e_e06008f921b7.slice/crio-075490582e1b4f6f85389dc5fe4b5d92a4cf927d312e0464f2634e784ab95481 WatchSource:0}: Error finding container 075490582e1b4f6f85389dc5fe4b5d92a4cf927d312e0464f2634e784ab95481: Status 404 returned error can't find the container with id 075490582e1b4f6f85389dc5fe4b5d92a4cf927d312e0464f2634e784ab95481 Dec 03 22:08:22 crc kubenswrapper[4715]: I1203 22:08:22.355042 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-rnccs"] Dec 03 22:08:22 crc kubenswrapper[4715]: W1203 22:08:22.360026 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbec62745_cf3a_4bc6_9fc6_34c8a56301ce.slice/crio-f100b4b3b604ce521bf19ddc1f2d8715df607c22ce18a681fd9afdc9a05b4559 WatchSource:0}: Error finding container f100b4b3b604ce521bf19ddc1f2d8715df607c22ce18a681fd9afdc9a05b4559: Status 404 returned error can't find the container with id f100b4b3b604ce521bf19ddc1f2d8715df607c22ce18a681fd9afdc9a05b4559 Dec 03 22:08:22 crc kubenswrapper[4715]: I1203 22:08:22.958094 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rnccs" event={"ID":"bec62745-cf3a-4bc6-9fc6-34c8a56301ce","Type":"ContainerStarted","Data":"f100b4b3b604ce521bf19ddc1f2d8715df607c22ce18a681fd9afdc9a05b4559"} Dec 03 22:08:22 crc kubenswrapper[4715]: I1203 22:08:22.960561 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-4bvpr" event={"ID":"ce5706ce-67d9-4e61-8051-03113e8c3ac4","Type":"ContainerStarted","Data":"bc8c36f847e9bbded00196e5e0311e106f433e851d3ef1cf241f8721895df1af"} Dec 03 22:08:22 crc kubenswrapper[4715]: I1203 22:08:22.962198 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" event={"ID":"30ac245b-49dd-48b9-878e-e06008f921b7","Type":"ContainerStarted","Data":"075490582e1b4f6f85389dc5fe4b5d92a4cf927d312e0464f2634e784ab95481"} Dec 03 22:08:24 crc kubenswrapper[4715]: I1203 22:08:24.974203 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" event={"ID":"30ac245b-49dd-48b9-878e-e06008f921b7","Type":"ContainerStarted","Data":"9d767d32d83943604598560706a099c5d187722a42f8c4bba307ac1341f95ec5"} Dec 03 22:08:24 crc kubenswrapper[4715]: I1203 22:08:24.975842 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" Dec 03 22:08:24 crc kubenswrapper[4715]: I1203 22:08:24.978048 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-rnccs" event={"ID":"bec62745-cf3a-4bc6-9fc6-34c8a56301ce","Type":"ContainerStarted","Data":"7562bf042b22de205fa3075ca3a8ca9d1feb89860356ac0cf3408df4df248853"} Dec 03 22:08:24 crc kubenswrapper[4715]: I1203 22:08:24.992156 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" podStartSLOduration=1.669702483 podStartE2EDuration="3.992129928s" podCreationTimestamp="2025-12-03 22:08:21 +0000 UTC" firstStartedPulling="2025-12-03 22:08:22.287344623 +0000 UTC m=+1599.030055258" lastFinishedPulling="2025-12-03 22:08:24.609772108 +0000 UTC m=+1601.352482703" observedRunningTime="2025-12-03 22:08:24.98810454 +0000 UTC m=+1601.730815175" watchObservedRunningTime="2025-12-03 22:08:24.992129928 +0000 UTC m=+1601.734840553" Dec 03 22:08:25 crc kubenswrapper[4715]: I1203 22:08:25.013530 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-rnccs" podStartSLOduration=1.686544773 podStartE2EDuration="4.013484209s" podCreationTimestamp="2025-12-03 22:08:21 +0000 UTC" firstStartedPulling="2025-12-03 22:08:22.362379419 +0000 UTC m=+1599.105090014" lastFinishedPulling="2025-12-03 22:08:24.689318855 +0000 UTC m=+1601.432029450" observedRunningTime="2025-12-03 22:08:25.005900886 +0000 UTC m=+1601.748611491" watchObservedRunningTime="2025-12-03 22:08:25.013484209 +0000 UTC m=+1601.756194834" Dec 03 22:08:25 crc kubenswrapper[4715]: I1203 22:08:25.987043 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-4bvpr" event={"ID":"ce5706ce-67d9-4e61-8051-03113e8c3ac4","Type":"ContainerStarted","Data":"a5445381e9883b5ac860c3581661db314bfc46b691afe789790e9d67ca727f27"} Dec 03 22:08:26 crc kubenswrapper[4715]: I1203 22:08:26.010084 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-4bvpr" podStartSLOduration=1.587054884 podStartE2EDuration="5.010064666s" podCreationTimestamp="2025-12-03 22:08:21 +0000 UTC" firstStartedPulling="2025-12-03 22:08:22.222099009 +0000 UTC m=+1598.964809604" lastFinishedPulling="2025-12-03 22:08:25.645108791 +0000 UTC m=+1602.387819386" observedRunningTime="2025-12-03 22:08:26.008194496 +0000 UTC m=+1602.750905131" watchObservedRunningTime="2025-12-03 22:08:26.010064666 +0000 UTC m=+1602.752775271" Dec 03 22:08:26 crc kubenswrapper[4715]: I1203 22:08:26.635789 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:08:26 crc kubenswrapper[4715]: E1203 22:08:26.636189 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.604609 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8j8wk"] Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.605690 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovn-controller" containerID="cri-o://b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62" gracePeriod=30 Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.606119 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="northd" containerID="cri-o://afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53" gracePeriod=30 Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.606129 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="sbdb" containerID="cri-o://ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd" gracePeriod=30 Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.606251 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="nbdb" containerID="cri-o://63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f" gracePeriod=30 Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.606403 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kube-rbac-proxy-node" containerID="cri-o://fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da" gracePeriod=30 Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.606464 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5" gracePeriod=30 Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.606546 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovn-acl-logging" containerID="cri-o://7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6" gracePeriod=30 Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.683618 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" containerID="cri-o://60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" gracePeriod=30 Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.984554 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-dgskj" Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.990347 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/3.log" Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.993101 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovn-acl-logging/0.log" Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.993733 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovn-controller/0.log" Dec 03 22:08:31 crc kubenswrapper[4715]: I1203 22:08:31.994417 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.032420 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovnkube-controller/3.log" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.035008 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovn-acl-logging/0.log" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.037285 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-8j8wk_ba9d6059-9d98-4816-8fad-9e430e516d70/ovn-controller/0.log" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.038942 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" exitCode=0 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039056 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039061 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd" exitCode=0 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039307 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f" exitCode=0 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039374 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53" exitCode=0 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039435 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5" exitCode=0 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039487 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da" exitCode=0 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039569 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6" exitCode=143 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039624 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerID="b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62" exitCode=143 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039013 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039876 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.039949 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040004 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040072 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040135 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040197 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040280 4715 scope.go:117] "RemoveContainer" containerID="60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040282 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040383 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040395 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040403 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040409 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040414 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040421 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040427 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040445 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040464 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040471 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040477 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040482 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040488 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040493 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040543 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040550 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040555 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040560 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040568 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040577 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040585 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040591 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040596 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040602 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040607 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040613 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040618 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040623 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040629 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040636 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-8j8wk" event={"ID":"ba9d6059-9d98-4816-8fad-9e430e516d70","Type":"ContainerDied","Data":"abb624dad13fd84bb42649176ed3e7cc27671d5a70a124f57d0baf90db00b41f"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040647 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040654 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040660 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040667 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040673 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040680 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040686 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040691 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040696 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.040701 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.044372 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/2.log" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.044834 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/1.log" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.044884 4715 generic.go:334] "Generic (PLEG): container finished" podID="9174d88f-f7f0-4b0c-942f-39d027b97e34" containerID="17360e95928ea8e902db39b173fee1daa0664875d571c8b1a6a2325810f6abbe" exitCode=2 Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.044921 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2wcn9" event={"ID":"9174d88f-f7f0-4b0c-942f-39d027b97e34","Type":"ContainerDied","Data":"17360e95928ea8e902db39b173fee1daa0664875d571c8b1a6a2325810f6abbe"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.044949 4715 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72"} Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.045536 4715 scope.go:117] "RemoveContainer" containerID="17360e95928ea8e902db39b173fee1daa0664875d571c8b1a6a2325810f6abbe" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.068837 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.070968 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f8cr7"] Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.071406 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.071478 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.071633 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.071690 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.071742 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovn-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.071807 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovn-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.071875 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kubecfg-setup" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.071930 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kubecfg-setup" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.071982 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="nbdb" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072029 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="nbdb" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.072089 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="sbdb" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072144 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="sbdb" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.072212 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovn-acl-logging" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072269 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovn-acl-logging" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.072322 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="northd" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072369 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="northd" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.072424 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kube-rbac-proxy-node" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072475 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kube-rbac-proxy-node" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.072560 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072619 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.072671 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072721 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072883 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072941 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="sbdb" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.072993 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovn-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073046 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073108 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073167 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073263 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kube-rbac-proxy-node" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073315 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073367 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovn-acl-logging" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073420 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="northd" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073474 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="nbdb" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.073679 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073755 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.073815 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.073881 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.074036 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" containerName="ovnkube-controller" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.078996 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.091763 4715 scope.go:117] "RemoveContainer" containerID="ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.096690 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-script-lib\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097102 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097252 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-bin\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097333 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097431 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-ovn-kubernetes\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097553 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-etc-openvswitch\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097583 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-openvswitch\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097621 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-env-overrides\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097633 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097646 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-kubelet\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097677 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097706 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpm4j\" (UniqueName: \"kubernetes.io/projected/ba9d6059-9d98-4816-8fad-9e430e516d70-kube-api-access-xpm4j\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097737 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-var-lib-openvswitch\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097674 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097758 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-netns\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097706 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097739 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097794 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-config\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097829 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-netd\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097858 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-systemd-units\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097880 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-systemd\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097904 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba9d6059-9d98-4816-8fad-9e430e516d70-ovn-node-metrics-cert\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097934 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-ovn\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097959 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.097963 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-slash\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098030 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-node-log\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098029 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-slash" (OuterVolumeSpecName: "host-slash") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098068 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098090 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-log-socket\") pod \"ba9d6059-9d98-4816-8fad-9e430e516d70\" (UID: \"ba9d6059-9d98-4816-8fad-9e430e516d70\") " Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098086 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098114 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098184 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098211 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-node-log" (OuterVolumeSpecName: "node-log") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098404 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-run-netns\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098431 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-node-log\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098460 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-env-overrides\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098481 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-systemd\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098547 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-slash\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098560 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098570 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-cni-netd\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098606 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-log-socket" (OuterVolumeSpecName: "log-socket") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovnkube-config\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098682 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovnkube-script-lib\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098696 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-ovn\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.098907 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-var-lib-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-cni-bin\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099205 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-systemd-units\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099280 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-etc-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099329 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-kubelet\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099358 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-run-ovn-kubernetes\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099409 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-log-socket\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099436 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovn-node-metrics-cert\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099467 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099517 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099550 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk72w\" (UniqueName: \"kubernetes.io/projected/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-kube-api-access-hk72w\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099634 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099748 4715 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099815 4715 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099874 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.099977 4715 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100049 4715 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100111 4715 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100192 4715 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100249 4715 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100379 4715 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100456 4715 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100578 4715 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100727 4715 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ba9d6059-9d98-4816-8fad-9e430e516d70-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100813 4715 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.100888 4715 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.101163 4715 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.101238 4715 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.103461 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba9d6059-9d98-4816-8fad-9e430e516d70-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.105935 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba9d6059-9d98-4816-8fad-9e430e516d70-kube-api-access-xpm4j" (OuterVolumeSpecName: "kube-api-access-xpm4j") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "kube-api-access-xpm4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.118622 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "ba9d6059-9d98-4816-8fad-9e430e516d70" (UID: "ba9d6059-9d98-4816-8fad-9e430e516d70"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.122599 4715 scope.go:117] "RemoveContainer" containerID="63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.138992 4715 scope.go:117] "RemoveContainer" containerID="afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.158688 4715 scope.go:117] "RemoveContainer" containerID="c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.172926 4715 scope.go:117] "RemoveContainer" containerID="fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.185617 4715 scope.go:117] "RemoveContainer" containerID="7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.199610 4715 scope.go:117] "RemoveContainer" containerID="b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.204269 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-ovn\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.205588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-ovn\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.205819 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-var-lib-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.205918 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-var-lib-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.205925 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-cni-bin\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.205963 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-cni-bin\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.205993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-systemd-units\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206094 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-etc-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206133 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-systemd-units\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-kubelet\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206179 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-etc-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206211 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-run-ovn-kubernetes\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206277 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-kubelet\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206274 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-log-socket\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206313 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-run-ovn-kubernetes\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206338 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovn-node-metrics-cert\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206404 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206459 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206546 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk72w\" (UniqueName: \"kubernetes.io/projected/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-kube-api-access-hk72w\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206612 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-run-netns\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-node-log\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206735 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-env-overrides\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206793 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-systemd\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206831 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-run-netns\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206841 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-openvswitch\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206875 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-slash\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206924 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-run-systemd\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206930 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-cni-netd\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206960 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-node-log\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206973 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovnkube-config\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.207023 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovnkube-script-lib\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.207114 4715 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.207151 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpm4j\" (UniqueName: \"kubernetes.io/projected/ba9d6059-9d98-4816-8fad-9e430e516d70-kube-api-access-xpm4j\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206354 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-log-socket\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.207186 4715 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/ba9d6059-9d98-4816-8fad-9e430e516d70-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.206800 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.207216 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ba9d6059-9d98-4816-8fad-9e430e516d70-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.207291 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-slash\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.207336 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-host-cni-netd\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.207933 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-env-overrides\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.208134 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovnkube-config\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.209004 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovnkube-script-lib\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.210089 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-ovn-node-metrics-cert\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.219332 4715 scope.go:117] "RemoveContainer" containerID="5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.224084 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk72w\" (UniqueName: \"kubernetes.io/projected/7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9-kube-api-access-hk72w\") pod \"ovnkube-node-f8cr7\" (UID: \"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.235275 4715 scope.go:117] "RemoveContainer" containerID="60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.235750 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": container with ID starting with 60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa not found: ID does not exist" containerID="60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.235786 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} err="failed to get container status \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": rpc error: code = NotFound desc = could not find container \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": container with ID starting with 60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.235808 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.236175 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": container with ID starting with c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e not found: ID does not exist" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.236208 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} err="failed to get container status \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": rpc error: code = NotFound desc = could not find container \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": container with ID starting with c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.236235 4715 scope.go:117] "RemoveContainer" containerID="ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.236654 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": container with ID starting with ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd not found: ID does not exist" containerID="ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.236678 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} err="failed to get container status \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": rpc error: code = NotFound desc = could not find container \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": container with ID starting with ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.236694 4715 scope.go:117] "RemoveContainer" containerID="63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.237023 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": container with ID starting with 63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f not found: ID does not exist" containerID="63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.237077 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} err="failed to get container status \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": rpc error: code = NotFound desc = could not find container \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": container with ID starting with 63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.237110 4715 scope.go:117] "RemoveContainer" containerID="afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.237415 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": container with ID starting with afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53 not found: ID does not exist" containerID="afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.237442 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} err="failed to get container status \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": rpc error: code = NotFound desc = could not find container \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": container with ID starting with afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.237460 4715 scope.go:117] "RemoveContainer" containerID="c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.237869 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": container with ID starting with c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5 not found: ID does not exist" containerID="c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.237902 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} err="failed to get container status \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": rpc error: code = NotFound desc = could not find container \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": container with ID starting with c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.237922 4715 scope.go:117] "RemoveContainer" containerID="fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.238241 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": container with ID starting with fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da not found: ID does not exist" containerID="fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.238272 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} err="failed to get container status \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": rpc error: code = NotFound desc = could not find container \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": container with ID starting with fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.238289 4715 scope.go:117] "RemoveContainer" containerID="7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.238809 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": container with ID starting with 7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6 not found: ID does not exist" containerID="7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.238899 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} err="failed to get container status \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": rpc error: code = NotFound desc = could not find container \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": container with ID starting with 7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.238972 4715 scope.go:117] "RemoveContainer" containerID="b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.239430 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": container with ID starting with b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62 not found: ID does not exist" containerID="b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.239544 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} err="failed to get container status \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": rpc error: code = NotFound desc = could not find container \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": container with ID starting with b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.239608 4715 scope.go:117] "RemoveContainer" containerID="5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7" Dec 03 22:08:32 crc kubenswrapper[4715]: E1203 22:08:32.239929 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": container with ID starting with 5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7 not found: ID does not exist" containerID="5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.239959 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} err="failed to get container status \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": rpc error: code = NotFound desc = could not find container \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": container with ID starting with 5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.239975 4715 scope.go:117] "RemoveContainer" containerID="60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.240340 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} err="failed to get container status \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": rpc error: code = NotFound desc = could not find container \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": container with ID starting with 60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.240384 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.240732 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} err="failed to get container status \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": rpc error: code = NotFound desc = could not find container \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": container with ID starting with c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.240754 4715 scope.go:117] "RemoveContainer" containerID="ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.241014 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} err="failed to get container status \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": rpc error: code = NotFound desc = could not find container \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": container with ID starting with ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.241040 4715 scope.go:117] "RemoveContainer" containerID="63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.241281 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} err="failed to get container status \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": rpc error: code = NotFound desc = could not find container \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": container with ID starting with 63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.241323 4715 scope.go:117] "RemoveContainer" containerID="afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.241696 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} err="failed to get container status \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": rpc error: code = NotFound desc = could not find container \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": container with ID starting with afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.241739 4715 scope.go:117] "RemoveContainer" containerID="c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.242147 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} err="failed to get container status \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": rpc error: code = NotFound desc = could not find container \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": container with ID starting with c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.242217 4715 scope.go:117] "RemoveContainer" containerID="fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.242543 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} err="failed to get container status \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": rpc error: code = NotFound desc = could not find container \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": container with ID starting with fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.242588 4715 scope.go:117] "RemoveContainer" containerID="7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.243004 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} err="failed to get container status \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": rpc error: code = NotFound desc = could not find container \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": container with ID starting with 7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.243073 4715 scope.go:117] "RemoveContainer" containerID="b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.243366 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} err="failed to get container status \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": rpc error: code = NotFound desc = could not find container \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": container with ID starting with b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.243388 4715 scope.go:117] "RemoveContainer" containerID="5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.243716 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} err="failed to get container status \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": rpc error: code = NotFound desc = could not find container \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": container with ID starting with 5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.243757 4715 scope.go:117] "RemoveContainer" containerID="60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.244228 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} err="failed to get container status \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": rpc error: code = NotFound desc = could not find container \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": container with ID starting with 60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.244260 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.244524 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} err="failed to get container status \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": rpc error: code = NotFound desc = could not find container \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": container with ID starting with c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.244548 4715 scope.go:117] "RemoveContainer" containerID="ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.244834 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} err="failed to get container status \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": rpc error: code = NotFound desc = could not find container \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": container with ID starting with ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.244863 4715 scope.go:117] "RemoveContainer" containerID="63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.245147 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} err="failed to get container status \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": rpc error: code = NotFound desc = could not find container \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": container with ID starting with 63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.245168 4715 scope.go:117] "RemoveContainer" containerID="afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.245482 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} err="failed to get container status \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": rpc error: code = NotFound desc = could not find container \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": container with ID starting with afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.245546 4715 scope.go:117] "RemoveContainer" containerID="c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.245844 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} err="failed to get container status \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": rpc error: code = NotFound desc = could not find container \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": container with ID starting with c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.245886 4715 scope.go:117] "RemoveContainer" containerID="fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.246179 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} err="failed to get container status \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": rpc error: code = NotFound desc = could not find container \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": container with ID starting with fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.246217 4715 scope.go:117] "RemoveContainer" containerID="7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.246578 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} err="failed to get container status \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": rpc error: code = NotFound desc = could not find container \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": container with ID starting with 7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.246600 4715 scope.go:117] "RemoveContainer" containerID="b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.246867 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} err="failed to get container status \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": rpc error: code = NotFound desc = could not find container \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": container with ID starting with b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.246909 4715 scope.go:117] "RemoveContainer" containerID="5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.247190 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} err="failed to get container status \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": rpc error: code = NotFound desc = could not find container \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": container with ID starting with 5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.247223 4715 scope.go:117] "RemoveContainer" containerID="60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.247515 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} err="failed to get container status \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": rpc error: code = NotFound desc = could not find container \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": container with ID starting with 60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.247546 4715 scope.go:117] "RemoveContainer" containerID="c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.247812 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e"} err="failed to get container status \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": rpc error: code = NotFound desc = could not find container \"c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e\": container with ID starting with c29e84b5c89fa25cfafa46a95d0a861552613020ba267c2a13a5c697e4a5782e not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.247842 4715 scope.go:117] "RemoveContainer" containerID="ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.248146 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd"} err="failed to get container status \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": rpc error: code = NotFound desc = could not find container \"ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd\": container with ID starting with ecff895becb94b79d7058a9d4251692359e1fc7fa7327d0fcdf116a0063f1acd not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.248183 4715 scope.go:117] "RemoveContainer" containerID="63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.248450 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f"} err="failed to get container status \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": rpc error: code = NotFound desc = could not find container \"63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f\": container with ID starting with 63bfbaa4f9d2a1c548218b77fcdf3bf2d269c7649c275ff07d03b1996c92a90f not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.248473 4715 scope.go:117] "RemoveContainer" containerID="afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.248737 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53"} err="failed to get container status \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": rpc error: code = NotFound desc = could not find container \"afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53\": container with ID starting with afebc0efbf2223c9d216e048917c5fbbe2c03c9b4c330c9fedf292cbe7e3cb53 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.248778 4715 scope.go:117] "RemoveContainer" containerID="c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.249124 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5"} err="failed to get container status \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": rpc error: code = NotFound desc = could not find container \"c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5\": container with ID starting with c74a6e982c582fec0b8947380c1b09c3840cd073be7ac933983f2dc2010e2ca5 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.249156 4715 scope.go:117] "RemoveContainer" containerID="fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.249499 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da"} err="failed to get container status \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": rpc error: code = NotFound desc = could not find container \"fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da\": container with ID starting with fab353178420f7f1d4958dbc2970bb31250911c86a26c8a5e8a3398ea524c2da not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.249566 4715 scope.go:117] "RemoveContainer" containerID="7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.249880 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6"} err="failed to get container status \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": rpc error: code = NotFound desc = could not find container \"7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6\": container with ID starting with 7209e64d6d1e672c297abacc512cafb6e697732a406ad09f4f1493f4b9ec54d6 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.249902 4715 scope.go:117] "RemoveContainer" containerID="b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.250178 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62"} err="failed to get container status \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": rpc error: code = NotFound desc = could not find container \"b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62\": container with ID starting with b95d5ce8f6aa4e8febef1f9e5236ff5b208286c1a126d208b8ded6623a28fe62 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.250220 4715 scope.go:117] "RemoveContainer" containerID="5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.250547 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7"} err="failed to get container status \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": rpc error: code = NotFound desc = could not find container \"5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7\": container with ID starting with 5f300fe46af987c0c76f073c2c37c990e017e361d33c1bf9cec433db9013f7c7 not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.250572 4715 scope.go:117] "RemoveContainer" containerID="60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.250851 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa"} err="failed to get container status \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": rpc error: code = NotFound desc = could not find container \"60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa\": container with ID starting with 60deadf37fb67eafc76db42f6d914c8f123d750461c9dec5e71459b7992b28aa not found: ID does not exist" Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.396426 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8j8wk"] Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.403078 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-8j8wk"] Dec 03 22:08:32 crc kubenswrapper[4715]: I1203 22:08:32.429943 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:32 crc kubenswrapper[4715]: W1203 22:08:32.460126 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dd0c8d2_4d0a_417d_810a_92ca72ec6dd9.slice/crio-2bfa26967bcc3386498bff88444728e5170a9bf7d7a616b48bbfcd8cc4f67a0e WatchSource:0}: Error finding container 2bfa26967bcc3386498bff88444728e5170a9bf7d7a616b48bbfcd8cc4f67a0e: Status 404 returned error can't find the container with id 2bfa26967bcc3386498bff88444728e5170a9bf7d7a616b48bbfcd8cc4f67a0e Dec 03 22:08:33 crc kubenswrapper[4715]: I1203 22:08:33.056090 4715 generic.go:334] "Generic (PLEG): container finished" podID="7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9" containerID="bad407191442fe26f70be383ccb50eb6360b3d17615547193cb39b41f8905b81" exitCode=0 Dec 03 22:08:33 crc kubenswrapper[4715]: I1203 22:08:33.056192 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerDied","Data":"bad407191442fe26f70be383ccb50eb6360b3d17615547193cb39b41f8905b81"} Dec 03 22:08:33 crc kubenswrapper[4715]: I1203 22:08:33.056846 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"2bfa26967bcc3386498bff88444728e5170a9bf7d7a616b48bbfcd8cc4f67a0e"} Dec 03 22:08:33 crc kubenswrapper[4715]: I1203 22:08:33.064826 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/2.log" Dec 03 22:08:33 crc kubenswrapper[4715]: I1203 22:08:33.065660 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/1.log" Dec 03 22:08:33 crc kubenswrapper[4715]: I1203 22:08:33.065753 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2wcn9" event={"ID":"9174d88f-f7f0-4b0c-942f-39d027b97e34","Type":"ContainerStarted","Data":"71af92e9b33fb13f73c2b67e7e9eef96b78570a0347dd6406d08d774671238ce"} Dec 03 22:08:33 crc kubenswrapper[4715]: I1203 22:08:33.640793 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba9d6059-9d98-4816-8fad-9e430e516d70" path="/var/lib/kubelet/pods/ba9d6059-9d98-4816-8fad-9e430e516d70/volumes" Dec 03 22:08:34 crc kubenswrapper[4715]: I1203 22:08:34.102737 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"0f33acacee65ad1340bf4d98682f91038f4713d0683427bb6232ee93d90d40a4"} Dec 03 22:08:34 crc kubenswrapper[4715]: I1203 22:08:34.102826 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"a572cab5821c3bda1657449e3466f10b73870ec0fb1efd893ec833912397cb67"} Dec 03 22:08:34 crc kubenswrapper[4715]: I1203 22:08:34.102851 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"ef129555239c8905484038a2e19ade5041e6becd2189821df61d6d209ee77ece"} Dec 03 22:08:34 crc kubenswrapper[4715]: I1203 22:08:34.102871 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"fef99ff634b0c890183b1fb64e4337145ea061c27717ea4ade13212cf95421e5"} Dec 03 22:08:34 crc kubenswrapper[4715]: I1203 22:08:34.102890 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"0b22e4156bc849106b7ab3ed06bfc96b1894f1ed631d6ad43419bc3a7083ec09"} Dec 03 22:08:34 crc kubenswrapper[4715]: I1203 22:08:34.102908 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"5cd36c1d80b2f7a46a32e2b342b70fe55ec285087db90d81a84c1a4e292a101d"} Dec 03 22:08:37 crc kubenswrapper[4715]: I1203 22:08:37.129490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"b041372f63fcac0121baaecdcf431568041aef1ca18fc38dadb19a25d419650d"} Dec 03 22:08:38 crc kubenswrapper[4715]: I1203 22:08:38.635052 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:08:38 crc kubenswrapper[4715]: E1203 22:08:38.636184 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:08:40 crc kubenswrapper[4715]: I1203 22:08:40.155159 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" event={"ID":"7dd0c8d2-4d0a-417d-810a-92ca72ec6dd9","Type":"ContainerStarted","Data":"86dbcded91d987fd0cbcc4898d4e03e44ceed5780cd0edc089618fce39fb4853"} Dec 03 22:08:40 crc kubenswrapper[4715]: I1203 22:08:40.155665 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:40 crc kubenswrapper[4715]: I1203 22:08:40.155681 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:40 crc kubenswrapper[4715]: I1203 22:08:40.155690 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:40 crc kubenswrapper[4715]: I1203 22:08:40.187055 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:40 crc kubenswrapper[4715]: I1203 22:08:40.189735 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:08:40 crc kubenswrapper[4715]: I1203 22:08:40.199479 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" podStartSLOduration=8.199454312 podStartE2EDuration="8.199454312s" podCreationTimestamp="2025-12-03 22:08:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:08:40.195808955 +0000 UTC m=+1616.938519580" watchObservedRunningTime="2025-12-03 22:08:40.199454312 +0000 UTC m=+1616.942164907" Dec 03 22:08:44 crc kubenswrapper[4715]: I1203 22:08:44.242737 4715 scope.go:117] "RemoveContainer" containerID="d4c4e7056b1c7731163dc82a7257e0d729cec39d08fe01800c85197a912e3f72" Dec 03 22:08:45 crc kubenswrapper[4715]: I1203 22:08:45.192403 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2wcn9_9174d88f-f7f0-4b0c-942f-39d027b97e34/kube-multus/2.log" Dec 03 22:08:50 crc kubenswrapper[4715]: I1203 22:08:50.634933 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:08:50 crc kubenswrapper[4715]: E1203 22:08:50.635637 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:09:02 crc kubenswrapper[4715]: I1203 22:09:02.463945 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f8cr7" Dec 03 22:09:05 crc kubenswrapper[4715]: I1203 22:09:05.634753 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:09:05 crc kubenswrapper[4715]: E1203 22:09:05.636126 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:09:16 crc kubenswrapper[4715]: I1203 22:09:16.634670 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:09:16 crc kubenswrapper[4715]: E1203 22:09:16.635563 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.407586 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr"] Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.410120 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.413167 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.416128 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr"] Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.491013 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgqb4\" (UniqueName: \"kubernetes.io/projected/d93c04f8-0649-47f4-b953-4dba2455c11e-kube-api-access-bgqb4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.491067 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.491113 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.592641 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgqb4\" (UniqueName: \"kubernetes.io/projected/d93c04f8-0649-47f4-b953-4dba2455c11e-kube-api-access-bgqb4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.592697 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.592749 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.593740 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.593485 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.628083 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgqb4\" (UniqueName: \"kubernetes.io/projected/d93c04f8-0649-47f4-b953-4dba2455c11e-kube-api-access-bgqb4\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.726303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:18 crc kubenswrapper[4715]: I1203 22:09:18.973996 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr"] Dec 03 22:09:19 crc kubenswrapper[4715]: I1203 22:09:19.413815 4715 generic.go:334] "Generic (PLEG): container finished" podID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerID="0df190568b85619508881daa862b881798baddf8efeff52ddd41c213c4c01d57" exitCode=0 Dec 03 22:09:19 crc kubenswrapper[4715]: I1203 22:09:19.413901 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" event={"ID":"d93c04f8-0649-47f4-b953-4dba2455c11e","Type":"ContainerDied","Data":"0df190568b85619508881daa862b881798baddf8efeff52ddd41c213c4c01d57"} Dec 03 22:09:19 crc kubenswrapper[4715]: I1203 22:09:19.413982 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" event={"ID":"d93c04f8-0649-47f4-b953-4dba2455c11e","Type":"ContainerStarted","Data":"ee706925a173887e71e1a81c56b36aedc23228bb9e207db0776915b880beecf2"} Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.756850 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6j7jh"] Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.763597 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.767793 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6j7jh"] Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.825918 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-catalog-content\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.825998 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgxr7\" (UniqueName: \"kubernetes.io/projected/5cedcea1-389c-4a34-bebb-e192f37d92a3-kube-api-access-kgxr7\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.826068 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-utilities\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.927900 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-utilities\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.928024 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-catalog-content\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.928054 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgxr7\" (UniqueName: \"kubernetes.io/projected/5cedcea1-389c-4a34-bebb-e192f37d92a3-kube-api-access-kgxr7\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.928778 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-utilities\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.928807 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-catalog-content\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:20 crc kubenswrapper[4715]: I1203 22:09:20.951367 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgxr7\" (UniqueName: \"kubernetes.io/projected/5cedcea1-389c-4a34-bebb-e192f37d92a3-kube-api-access-kgxr7\") pod \"redhat-operators-6j7jh\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:21 crc kubenswrapper[4715]: I1203 22:09:21.098307 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:21 crc kubenswrapper[4715]: I1203 22:09:21.317595 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6j7jh"] Dec 03 22:09:21 crc kubenswrapper[4715]: I1203 22:09:21.426595 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6j7jh" event={"ID":"5cedcea1-389c-4a34-bebb-e192f37d92a3","Type":"ContainerStarted","Data":"0ef0ef02adb320b11b301e1c61a515fe64f8f23c909d50fa15d6ddbb4279684a"} Dec 03 22:09:22 crc kubenswrapper[4715]: I1203 22:09:22.434723 4715 generic.go:334] "Generic (PLEG): container finished" podID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerID="4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd" exitCode=0 Dec 03 22:09:22 crc kubenswrapper[4715]: I1203 22:09:22.434808 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6j7jh" event={"ID":"5cedcea1-389c-4a34-bebb-e192f37d92a3","Type":"ContainerDied","Data":"4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd"} Dec 03 22:09:24 crc kubenswrapper[4715]: I1203 22:09:24.449644 4715 generic.go:334] "Generic (PLEG): container finished" podID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerID="c04b6e793449256a4d7eae11124b3d22a66b20b08572cd1a81a3fba85baf72bd" exitCode=0 Dec 03 22:09:24 crc kubenswrapper[4715]: I1203 22:09:24.449731 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" event={"ID":"d93c04f8-0649-47f4-b953-4dba2455c11e","Type":"ContainerDied","Data":"c04b6e793449256a4d7eae11124b3d22a66b20b08572cd1a81a3fba85baf72bd"} Dec 03 22:09:25 crc kubenswrapper[4715]: I1203 22:09:25.459037 4715 generic.go:334] "Generic (PLEG): container finished" podID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerID="532ce434410c508bd9be4a5cf6b4b20d23a48a03e4d6cc99d39a7c78f0c7aa5e" exitCode=0 Dec 03 22:09:25 crc kubenswrapper[4715]: I1203 22:09:25.459140 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" event={"ID":"d93c04f8-0649-47f4-b953-4dba2455c11e","Type":"ContainerDied","Data":"532ce434410c508bd9be4a5cf6b4b20d23a48a03e4d6cc99d39a7c78f0c7aa5e"} Dec 03 22:09:26 crc kubenswrapper[4715]: I1203 22:09:26.772738 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:26 crc kubenswrapper[4715]: I1203 22:09:26.920974 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-bundle\") pod \"d93c04f8-0649-47f4-b953-4dba2455c11e\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " Dec 03 22:09:26 crc kubenswrapper[4715]: I1203 22:09:26.921082 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgqb4\" (UniqueName: \"kubernetes.io/projected/d93c04f8-0649-47f4-b953-4dba2455c11e-kube-api-access-bgqb4\") pod \"d93c04f8-0649-47f4-b953-4dba2455c11e\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " Dec 03 22:09:26 crc kubenswrapper[4715]: I1203 22:09:26.921166 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-util\") pod \"d93c04f8-0649-47f4-b953-4dba2455c11e\" (UID: \"d93c04f8-0649-47f4-b953-4dba2455c11e\") " Dec 03 22:09:26 crc kubenswrapper[4715]: I1203 22:09:26.924223 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-bundle" (OuterVolumeSpecName: "bundle") pod "d93c04f8-0649-47f4-b953-4dba2455c11e" (UID: "d93c04f8-0649-47f4-b953-4dba2455c11e"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:09:26 crc kubenswrapper[4715]: I1203 22:09:26.930924 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-util" (OuterVolumeSpecName: "util") pod "d93c04f8-0649-47f4-b953-4dba2455c11e" (UID: "d93c04f8-0649-47f4-b953-4dba2455c11e"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:09:26 crc kubenswrapper[4715]: I1203 22:09:26.932587 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d93c04f8-0649-47f4-b953-4dba2455c11e-kube-api-access-bgqb4" (OuterVolumeSpecName: "kube-api-access-bgqb4") pod "d93c04f8-0649-47f4-b953-4dba2455c11e" (UID: "d93c04f8-0649-47f4-b953-4dba2455c11e"). InnerVolumeSpecName "kube-api-access-bgqb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:09:27 crc kubenswrapper[4715]: I1203 22:09:27.022721 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-util\") on node \"crc\" DevicePath \"\"" Dec 03 22:09:27 crc kubenswrapper[4715]: I1203 22:09:27.022761 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d93c04f8-0649-47f4-b953-4dba2455c11e-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:09:27 crc kubenswrapper[4715]: I1203 22:09:27.022774 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgqb4\" (UniqueName: \"kubernetes.io/projected/d93c04f8-0649-47f4-b953-4dba2455c11e-kube-api-access-bgqb4\") on node \"crc\" DevicePath \"\"" Dec 03 22:09:27 crc kubenswrapper[4715]: I1203 22:09:27.475674 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6j7jh" event={"ID":"5cedcea1-389c-4a34-bebb-e192f37d92a3","Type":"ContainerStarted","Data":"fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125"} Dec 03 22:09:27 crc kubenswrapper[4715]: I1203 22:09:27.487801 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" event={"ID":"d93c04f8-0649-47f4-b953-4dba2455c11e","Type":"ContainerDied","Data":"ee706925a173887e71e1a81c56b36aedc23228bb9e207db0776915b880beecf2"} Dec 03 22:09:27 crc kubenswrapper[4715]: I1203 22:09:27.487878 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee706925a173887e71e1a81c56b36aedc23228bb9e207db0776915b880beecf2" Dec 03 22:09:27 crc kubenswrapper[4715]: I1203 22:09:27.487993 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.498842 4715 generic.go:334] "Generic (PLEG): container finished" podID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerID="fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125" exitCode=0 Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.498911 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6j7jh" event={"ID":"5cedcea1-389c-4a34-bebb-e192f37d92a3","Type":"ContainerDied","Data":"fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125"} Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.635037 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:09:28 crc kubenswrapper[4715]: E1203 22:09:28.635759 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.951538 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s"] Dec 03 22:09:28 crc kubenswrapper[4715]: E1203 22:09:28.951749 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerName="util" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.951759 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerName="util" Dec 03 22:09:28 crc kubenswrapper[4715]: E1203 22:09:28.951769 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerName="extract" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.951774 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerName="extract" Dec 03 22:09:28 crc kubenswrapper[4715]: E1203 22:09:28.951785 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerName="pull" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.951791 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerName="pull" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.951874 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d93c04f8-0649-47f4-b953-4dba2455c11e" containerName="extract" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.952214 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.954010 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.954195 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.954282 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-j7227" Dec 03 22:09:28 crc kubenswrapper[4715]: I1203 22:09:28.970590 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s"] Dec 03 22:09:29 crc kubenswrapper[4715]: I1203 22:09:29.048167 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkc2m\" (UniqueName: \"kubernetes.io/projected/7bff12a8-00c5-4a84-85f7-3a0055308b87-kube-api-access-mkc2m\") pod \"nmstate-operator-5b5b58f5c8-xht5s\" (UID: \"7bff12a8-00c5-4a84-85f7-3a0055308b87\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s" Dec 03 22:09:29 crc kubenswrapper[4715]: I1203 22:09:29.149061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkc2m\" (UniqueName: \"kubernetes.io/projected/7bff12a8-00c5-4a84-85f7-3a0055308b87-kube-api-access-mkc2m\") pod \"nmstate-operator-5b5b58f5c8-xht5s\" (UID: \"7bff12a8-00c5-4a84-85f7-3a0055308b87\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s" Dec 03 22:09:29 crc kubenswrapper[4715]: I1203 22:09:29.168646 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkc2m\" (UniqueName: \"kubernetes.io/projected/7bff12a8-00c5-4a84-85f7-3a0055308b87-kube-api-access-mkc2m\") pod \"nmstate-operator-5b5b58f5c8-xht5s\" (UID: \"7bff12a8-00c5-4a84-85f7-3a0055308b87\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s" Dec 03 22:09:29 crc kubenswrapper[4715]: I1203 22:09:29.268223 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s" Dec 03 22:09:29 crc kubenswrapper[4715]: I1203 22:09:29.508736 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s"] Dec 03 22:09:29 crc kubenswrapper[4715]: W1203 22:09:29.528818 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bff12a8_00c5_4a84_85f7_3a0055308b87.slice/crio-38eb750f0be53e1e0a15533d491159ce74b25e493d26b764314ef75bf9dc6083 WatchSource:0}: Error finding container 38eb750f0be53e1e0a15533d491159ce74b25e493d26b764314ef75bf9dc6083: Status 404 returned error can't find the container with id 38eb750f0be53e1e0a15533d491159ce74b25e493d26b764314ef75bf9dc6083 Dec 03 22:09:30 crc kubenswrapper[4715]: I1203 22:09:30.523133 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6j7jh" event={"ID":"5cedcea1-389c-4a34-bebb-e192f37d92a3","Type":"ContainerStarted","Data":"24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592"} Dec 03 22:09:30 crc kubenswrapper[4715]: I1203 22:09:30.525337 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s" event={"ID":"7bff12a8-00c5-4a84-85f7-3a0055308b87","Type":"ContainerStarted","Data":"38eb750f0be53e1e0a15533d491159ce74b25e493d26b764314ef75bf9dc6083"} Dec 03 22:09:30 crc kubenswrapper[4715]: I1203 22:09:30.545768 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6j7jh" podStartSLOduration=2.729104408 podStartE2EDuration="10.545738468s" podCreationTimestamp="2025-12-03 22:09:20 +0000 UTC" firstStartedPulling="2025-12-03 22:09:22.439682073 +0000 UTC m=+1659.182392678" lastFinishedPulling="2025-12-03 22:09:30.256316143 +0000 UTC m=+1666.999026738" observedRunningTime="2025-12-03 22:09:30.543192451 +0000 UTC m=+1667.285903056" watchObservedRunningTime="2025-12-03 22:09:30.545738468 +0000 UTC m=+1667.288449083" Dec 03 22:09:31 crc kubenswrapper[4715]: I1203 22:09:31.098626 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:31 crc kubenswrapper[4715]: I1203 22:09:31.098905 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:32 crc kubenswrapper[4715]: I1203 22:09:32.168566 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6j7jh" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="registry-server" probeResult="failure" output=< Dec 03 22:09:32 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 03 22:09:32 crc kubenswrapper[4715]: > Dec 03 22:09:32 crc kubenswrapper[4715]: I1203 22:09:32.544665 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s" event={"ID":"7bff12a8-00c5-4a84-85f7-3a0055308b87","Type":"ContainerStarted","Data":"eafe52bf26e4d61c957b877b487f1739beb37e7887fb9ae47705d0848c63224a"} Dec 03 22:09:32 crc kubenswrapper[4715]: I1203 22:09:32.567318 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-xht5s" podStartSLOduration=1.950083765 podStartE2EDuration="4.567299369s" podCreationTimestamp="2025-12-03 22:09:28 +0000 UTC" firstStartedPulling="2025-12-03 22:09:29.531701402 +0000 UTC m=+1666.274411997" lastFinishedPulling="2025-12-03 22:09:32.148917006 +0000 UTC m=+1668.891627601" observedRunningTime="2025-12-03 22:09:32.563747025 +0000 UTC m=+1669.306457620" watchObservedRunningTime="2025-12-03 22:09:32.567299369 +0000 UTC m=+1669.310009964" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.805428 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x"] Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.806784 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.812197 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-tsgfn" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.831811 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x"] Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.836565 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-vkmgw"] Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.837581 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.854389 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9"] Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.855124 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.859896 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.892844 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9"] Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.904451 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-dbus-socket\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.904570 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-ovs-socket\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.904792 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-nmstate-lock\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.904817 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfldm\" (UniqueName: \"kubernetes.io/projected/6d716f67-d1d1-44c0-b2d4-eeb60e73160f-kube-api-access-kfldm\") pod \"nmstate-metrics-7f946cbc9-9n62x\" (UID: \"6d716f67-d1d1-44c0-b2d4-eeb60e73160f\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.904863 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgb59\" (UniqueName: \"kubernetes.io/projected/be31b811-7008-4d9f-ac8b-4c454e4a8597-kube-api-access-zgb59\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.905025 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7l45\" (UniqueName: \"kubernetes.io/projected/bef574d3-8639-4a82-aa6e-323a58b9db40-kube-api-access-k7l45\") pod \"nmstate-webhook-5f6d4c5ccb-4d8n9\" (UID: \"bef574d3-8639-4a82-aa6e-323a58b9db40\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.905071 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bef574d3-8639-4a82-aa6e-323a58b9db40-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4d8n9\" (UID: \"bef574d3-8639-4a82-aa6e-323a58b9db40\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.956471 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr"] Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.957716 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.960327 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.961153 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-j5cwg" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.961323 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 22:09:38 crc kubenswrapper[4715]: I1203 22:09:38.975584 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr"] Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006319 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgb59\" (UniqueName: \"kubernetes.io/projected/be31b811-7008-4d9f-ac8b-4c454e4a8597-kube-api-access-zgb59\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006396 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47af66c3-ee08-4f79-9e8e-f9e1b3457971-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006426 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7l45\" (UniqueName: \"kubernetes.io/projected/bef574d3-8639-4a82-aa6e-323a58b9db40-kube-api-access-k7l45\") pod \"nmstate-webhook-5f6d4c5ccb-4d8n9\" (UID: \"bef574d3-8639-4a82-aa6e-323a58b9db40\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006444 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bef574d3-8639-4a82-aa6e-323a58b9db40-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4d8n9\" (UID: \"bef574d3-8639-4a82-aa6e-323a58b9db40\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006468 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-dbus-socket\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006522 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47af66c3-ee08-4f79-9e8e-f9e1b3457971-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006544 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl9ps\" (UniqueName: \"kubernetes.io/projected/47af66c3-ee08-4f79-9e8e-f9e1b3457971-kube-api-access-vl9ps\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006562 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-ovs-socket\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-nmstate-lock\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.006599 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfldm\" (UniqueName: \"kubernetes.io/projected/6d716f67-d1d1-44c0-b2d4-eeb60e73160f-kube-api-access-kfldm\") pod \"nmstate-metrics-7f946cbc9-9n62x\" (UID: \"6d716f67-d1d1-44c0-b2d4-eeb60e73160f\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.007168 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-ovs-socket\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.007201 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-nmstate-lock\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: E1203 22:09:39.007379 4715 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 03 22:09:39 crc kubenswrapper[4715]: E1203 22:09:39.007478 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bef574d3-8639-4a82-aa6e-323a58b9db40-tls-key-pair podName:bef574d3-8639-4a82-aa6e-323a58b9db40 nodeName:}" failed. No retries permitted until 2025-12-03 22:09:39.507451284 +0000 UTC m=+1676.250161879 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/bef574d3-8639-4a82-aa6e-323a58b9db40-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-4d8n9" (UID: "bef574d3-8639-4a82-aa6e-323a58b9db40") : secret "openshift-nmstate-webhook" not found Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.007527 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/be31b811-7008-4d9f-ac8b-4c454e4a8597-dbus-socket\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.028137 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfldm\" (UniqueName: \"kubernetes.io/projected/6d716f67-d1d1-44c0-b2d4-eeb60e73160f-kube-api-access-kfldm\") pod \"nmstate-metrics-7f946cbc9-9n62x\" (UID: \"6d716f67-d1d1-44c0-b2d4-eeb60e73160f\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.028189 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgb59\" (UniqueName: \"kubernetes.io/projected/be31b811-7008-4d9f-ac8b-4c454e4a8597-kube-api-access-zgb59\") pod \"nmstate-handler-vkmgw\" (UID: \"be31b811-7008-4d9f-ac8b-4c454e4a8597\") " pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.039327 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7l45\" (UniqueName: \"kubernetes.io/projected/bef574d3-8639-4a82-aa6e-323a58b9db40-kube-api-access-k7l45\") pod \"nmstate-webhook-5f6d4c5ccb-4d8n9\" (UID: \"bef574d3-8639-4a82-aa6e-323a58b9db40\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.107790 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47af66c3-ee08-4f79-9e8e-f9e1b3457971-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.108253 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47af66c3-ee08-4f79-9e8e-f9e1b3457971-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.108374 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl9ps\" (UniqueName: \"kubernetes.io/projected/47af66c3-ee08-4f79-9e8e-f9e1b3457971-kube-api-access-vl9ps\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.110185 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/47af66c3-ee08-4f79-9e8e-f9e1b3457971-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.114397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/47af66c3-ee08-4f79-9e8e-f9e1b3457971-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.127238 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.137442 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl9ps\" (UniqueName: \"kubernetes.io/projected/47af66c3-ee08-4f79-9e8e-f9e1b3457971-kube-api-access-vl9ps\") pod \"nmstate-console-plugin-7fbb5f6569-lkqpr\" (UID: \"47af66c3-ee08-4f79-9e8e-f9e1b3457971\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.154710 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.186410 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6b885bb987-84jfc"] Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.188585 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.200734 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b885bb987-84jfc"] Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.275453 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.311644 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-trusted-ca-bundle\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.311692 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1387107f-9ede-4678-b493-3d052cd752df-console-serving-cert\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.311727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1387107f-9ede-4678-b493-3d052cd752df-console-oauth-config\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.311749 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-console-config\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.311773 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr9zh\" (UniqueName: \"kubernetes.io/projected/1387107f-9ede-4678-b493-3d052cd752df-kube-api-access-lr9zh\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.311798 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-service-ca\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.311891 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-oauth-serving-cert\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.388604 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x"] Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.414227 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-oauth-serving-cert\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.414310 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-trusted-ca-bundle\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.414343 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1387107f-9ede-4678-b493-3d052cd752df-console-serving-cert\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.414377 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1387107f-9ede-4678-b493-3d052cd752df-console-oauth-config\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.414429 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-console-config\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.414463 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr9zh\" (UniqueName: \"kubernetes.io/projected/1387107f-9ede-4678-b493-3d052cd752df-kube-api-access-lr9zh\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.414544 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-service-ca\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.415409 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-oauth-serving-cert\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.415854 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-trusted-ca-bundle\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.417146 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-service-ca\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.417268 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/1387107f-9ede-4678-b493-3d052cd752df-console-config\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.421372 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/1387107f-9ede-4678-b493-3d052cd752df-console-serving-cert\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.421797 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/1387107f-9ede-4678-b493-3d052cd752df-console-oauth-config\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.434290 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr9zh\" (UniqueName: \"kubernetes.io/projected/1387107f-9ede-4678-b493-3d052cd752df-kube-api-access-lr9zh\") pod \"console-6b885bb987-84jfc\" (UID: \"1387107f-9ede-4678-b493-3d052cd752df\") " pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: W1203 22:09:39.512988 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47af66c3_ee08_4f79_9e8e_f9e1b3457971.slice/crio-64cd269daf9f90a124915b3093d52b9ecd201d55da9bea5ab6ad7a6fecbdaeb0 WatchSource:0}: Error finding container 64cd269daf9f90a124915b3093d52b9ecd201d55da9bea5ab6ad7a6fecbdaeb0: Status 404 returned error can't find the container with id 64cd269daf9f90a124915b3093d52b9ecd201d55da9bea5ab6ad7a6fecbdaeb0 Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.515468 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr"] Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.515999 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.516047 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bef574d3-8639-4a82-aa6e-323a58b9db40-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4d8n9\" (UID: \"bef574d3-8639-4a82-aa6e-323a58b9db40\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.520175 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/bef574d3-8639-4a82-aa6e-323a58b9db40-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-4d8n9\" (UID: \"bef574d3-8639-4a82-aa6e-323a58b9db40\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.597490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" event={"ID":"47af66c3-ee08-4f79-9e8e-f9e1b3457971","Type":"ContainerStarted","Data":"64cd269daf9f90a124915b3093d52b9ecd201d55da9bea5ab6ad7a6fecbdaeb0"} Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.604479 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" event={"ID":"6d716f67-d1d1-44c0-b2d4-eeb60e73160f","Type":"ContainerStarted","Data":"f1fb3b9247400972dae5364f10b9c182f1fa1dbf958779acb1d3fcbb672a7f89"} Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.607672 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vkmgw" event={"ID":"be31b811-7008-4d9f-ac8b-4c454e4a8597","Type":"ContainerStarted","Data":"5f134a98c05fc5d61535c290bfe4f39c98cd3e787ecd611d8b9e81366e903e10"} Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.735288 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6b885bb987-84jfc"] Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.777703 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:39 crc kubenswrapper[4715]: I1203 22:09:39.994469 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9"] Dec 03 22:09:40 crc kubenswrapper[4715]: W1203 22:09:40.005768 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbef574d3_8639_4a82_aa6e_323a58b9db40.slice/crio-409769839bf3867a8f8c86ba526379e17d62294a7b74be44f846a32d93d7f89f WatchSource:0}: Error finding container 409769839bf3867a8f8c86ba526379e17d62294a7b74be44f846a32d93d7f89f: Status 404 returned error can't find the container with id 409769839bf3867a8f8c86ba526379e17d62294a7b74be44f846a32d93d7f89f Dec 03 22:09:40 crc kubenswrapper[4715]: I1203 22:09:40.619252 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" event={"ID":"bef574d3-8639-4a82-aa6e-323a58b9db40","Type":"ContainerStarted","Data":"409769839bf3867a8f8c86ba526379e17d62294a7b74be44f846a32d93d7f89f"} Dec 03 22:09:40 crc kubenswrapper[4715]: I1203 22:09:40.621177 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b885bb987-84jfc" event={"ID":"1387107f-9ede-4678-b493-3d052cd752df","Type":"ContainerStarted","Data":"fbd05c5b6d0a8006a0eee2cb2d4f8699cfacf737d9770344ac05ec5b826378f4"} Dec 03 22:09:40 crc kubenswrapper[4715]: I1203 22:09:40.635031 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:09:40 crc kubenswrapper[4715]: E1203 22:09:40.635485 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:09:41 crc kubenswrapper[4715]: I1203 22:09:41.178163 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:41 crc kubenswrapper[4715]: I1203 22:09:41.261427 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:41 crc kubenswrapper[4715]: I1203 22:09:41.439058 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6j7jh"] Dec 03 22:09:42 crc kubenswrapper[4715]: I1203 22:09:42.636610 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6j7jh" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="registry-server" containerID="cri-o://24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592" gracePeriod=2 Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.527698 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.580067 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-catalog-content\") pod \"5cedcea1-389c-4a34-bebb-e192f37d92a3\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.580190 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgxr7\" (UniqueName: \"kubernetes.io/projected/5cedcea1-389c-4a34-bebb-e192f37d92a3-kube-api-access-kgxr7\") pod \"5cedcea1-389c-4a34-bebb-e192f37d92a3\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.580219 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-utilities\") pod \"5cedcea1-389c-4a34-bebb-e192f37d92a3\" (UID: \"5cedcea1-389c-4a34-bebb-e192f37d92a3\") " Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.581414 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-utilities" (OuterVolumeSpecName: "utilities") pod "5cedcea1-389c-4a34-bebb-e192f37d92a3" (UID: "5cedcea1-389c-4a34-bebb-e192f37d92a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.587793 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cedcea1-389c-4a34-bebb-e192f37d92a3-kube-api-access-kgxr7" (OuterVolumeSpecName: "kube-api-access-kgxr7") pod "5cedcea1-389c-4a34-bebb-e192f37d92a3" (UID: "5cedcea1-389c-4a34-bebb-e192f37d92a3"). InnerVolumeSpecName "kube-api-access-kgxr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.647682 4715 generic.go:334] "Generic (PLEG): container finished" podID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerID="24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592" exitCode=0 Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.647845 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6j7jh" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.660930 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6j7jh" event={"ID":"5cedcea1-389c-4a34-bebb-e192f37d92a3","Type":"ContainerDied","Data":"24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592"} Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.660983 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6j7jh" event={"ID":"5cedcea1-389c-4a34-bebb-e192f37d92a3","Type":"ContainerDied","Data":"0ef0ef02adb320b11b301e1c61a515fe64f8f23c909d50fa15d6ddbb4279684a"} Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.661000 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6b885bb987-84jfc" event={"ID":"1387107f-9ede-4678-b493-3d052cd752df","Type":"ContainerStarted","Data":"5a7dfd8e2bc51052549d6a55b81b7a8b43ea1db9dee6c36b30620c6b22258d05"} Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.661024 4715 scope.go:117] "RemoveContainer" containerID="24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.680601 4715 scope.go:117] "RemoveContainer" containerID="fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.682102 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgxr7\" (UniqueName: \"kubernetes.io/projected/5cedcea1-389c-4a34-bebb-e192f37d92a3-kube-api-access-kgxr7\") on node \"crc\" DevicePath \"\"" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.682144 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.712428 4715 scope.go:117] "RemoveContainer" containerID="4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.716125 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5cedcea1-389c-4a34-bebb-e192f37d92a3" (UID: "5cedcea1-389c-4a34-bebb-e192f37d92a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.723945 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6b885bb987-84jfc" podStartSLOduration=4.723922776 podStartE2EDuration="4.723922776s" podCreationTimestamp="2025-12-03 22:09:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:09:43.72144905 +0000 UTC m=+1680.464159645" watchObservedRunningTime="2025-12-03 22:09:43.723922776 +0000 UTC m=+1680.466633371" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.740678 4715 scope.go:117] "RemoveContainer" containerID="24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592" Dec 03 22:09:43 crc kubenswrapper[4715]: E1203 22:09:43.741170 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592\": container with ID starting with 24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592 not found: ID does not exist" containerID="24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.741203 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592"} err="failed to get container status \"24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592\": rpc error: code = NotFound desc = could not find container \"24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592\": container with ID starting with 24957c70c0a5bdaa9a7d2a9e263a6a53462c73ae7df6fb070d137fe61e5ee592 not found: ID does not exist" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.741225 4715 scope.go:117] "RemoveContainer" containerID="fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125" Dec 03 22:09:43 crc kubenswrapper[4715]: E1203 22:09:43.741625 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125\": container with ID starting with fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125 not found: ID does not exist" containerID="fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.741757 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125"} err="failed to get container status \"fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125\": rpc error: code = NotFound desc = could not find container \"fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125\": container with ID starting with fbe1429fa9b27d283b3e5f6c198febfee8fb8e1890dc3b24372eaa5387b25125 not found: ID does not exist" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.741800 4715 scope.go:117] "RemoveContainer" containerID="4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd" Dec 03 22:09:43 crc kubenswrapper[4715]: E1203 22:09:43.742460 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd\": container with ID starting with 4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd not found: ID does not exist" containerID="4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.742559 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd"} err="failed to get container status \"4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd\": rpc error: code = NotFound desc = could not find container \"4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd\": container with ID starting with 4599d60a8c6b59f3c2eafb3c9c966e468535d4a398d02139506e51c14806c4cd not found: ID does not exist" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.783727 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cedcea1-389c-4a34-bebb-e192f37d92a3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.977722 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6j7jh"] Dec 03 22:09:43 crc kubenswrapper[4715]: I1203 22:09:43.981657 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6j7jh"] Dec 03 22:09:45 crc kubenswrapper[4715]: I1203 22:09:45.648381 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" path="/var/lib/kubelet/pods/5cedcea1-389c-4a34-bebb-e192f37d92a3/volumes" Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.688686 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" event={"ID":"47af66c3-ee08-4f79-9e8e-f9e1b3457971","Type":"ContainerStarted","Data":"ca253a32e9edab03eb2dbfe5a56c6ca3a430b34826e06661ae9cece644283c4b"} Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.690164 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-vkmgw" event={"ID":"be31b811-7008-4d9f-ac8b-4c454e4a8597","Type":"ContainerStarted","Data":"287f968415c38f7e1343a3288d4e10de35040cae70069d115507312ecc36b264"} Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.690279 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.692748 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" event={"ID":"bef574d3-8639-4a82-aa6e-323a58b9db40","Type":"ContainerStarted","Data":"48b901ef053a6841ce098c567e427c41fa48df0675e421327fd8c6a656ee252b"} Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.692908 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.694597 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" event={"ID":"6d716f67-d1d1-44c0-b2d4-eeb60e73160f","Type":"ContainerStarted","Data":"1d49231dc74274c2b68dd23f40b1910df88c7b287a432a10e986a4e7ff1d7a9e"} Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.708317 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-lkqpr" podStartSLOduration=2.338700298 podStartE2EDuration="8.708293744s" podCreationTimestamp="2025-12-03 22:09:38 +0000 UTC" firstStartedPulling="2025-12-03 22:09:39.515759146 +0000 UTC m=+1676.258469741" lastFinishedPulling="2025-12-03 22:09:45.885352552 +0000 UTC m=+1682.628063187" observedRunningTime="2025-12-03 22:09:46.706017323 +0000 UTC m=+1683.448727968" watchObservedRunningTime="2025-12-03 22:09:46.708293744 +0000 UTC m=+1683.451004339" Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.727165 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" podStartSLOduration=2.839520653 podStartE2EDuration="8.72713804s" podCreationTimestamp="2025-12-03 22:09:38 +0000 UTC" firstStartedPulling="2025-12-03 22:09:40.008873438 +0000 UTC m=+1676.751584033" lastFinishedPulling="2025-12-03 22:09:45.896490825 +0000 UTC m=+1682.639201420" observedRunningTime="2025-12-03 22:09:46.725843795 +0000 UTC m=+1683.468554400" watchObservedRunningTime="2025-12-03 22:09:46.72713804 +0000 UTC m=+1683.469848665" Dec 03 22:09:46 crc kubenswrapper[4715]: I1203 22:09:46.744020 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-vkmgw" podStartSLOduration=2.039100365 podStartE2EDuration="8.743994744s" podCreationTimestamp="2025-12-03 22:09:38 +0000 UTC" firstStartedPulling="2025-12-03 22:09:39.191328399 +0000 UTC m=+1675.934038994" lastFinishedPulling="2025-12-03 22:09:45.896222778 +0000 UTC m=+1682.638933373" observedRunningTime="2025-12-03 22:09:46.741108568 +0000 UTC m=+1683.483819173" watchObservedRunningTime="2025-12-03 22:09:46.743994744 +0000 UTC m=+1683.486705349" Dec 03 22:09:48 crc kubenswrapper[4715]: I1203 22:09:48.712062 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" event={"ID":"6d716f67-d1d1-44c0-b2d4-eeb60e73160f","Type":"ContainerStarted","Data":"d9d1a0132064b04f71e242a3ec3db4652a17b70e3de56bdafba5b241b258ee6b"} Dec 03 22:09:48 crc kubenswrapper[4715]: I1203 22:09:48.744831 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9n62x" podStartSLOduration=1.907825856 podStartE2EDuration="10.744803098s" podCreationTimestamp="2025-12-03 22:09:38 +0000 UTC" firstStartedPulling="2025-12-03 22:09:39.409203499 +0000 UTC m=+1676.151914094" lastFinishedPulling="2025-12-03 22:09:48.246180741 +0000 UTC m=+1684.988891336" observedRunningTime="2025-12-03 22:09:48.736771166 +0000 UTC m=+1685.479481791" watchObservedRunningTime="2025-12-03 22:09:48.744803098 +0000 UTC m=+1685.487513733" Dec 03 22:09:49 crc kubenswrapper[4715]: I1203 22:09:49.516486 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:49 crc kubenswrapper[4715]: I1203 22:09:49.516629 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:49 crc kubenswrapper[4715]: I1203 22:09:49.534454 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:49 crc kubenswrapper[4715]: I1203 22:09:49.729644 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6b885bb987-84jfc" Dec 03 22:09:49 crc kubenswrapper[4715]: I1203 22:09:49.795364 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sppb5"] Dec 03 22:09:54 crc kubenswrapper[4715]: I1203 22:09:54.198452 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-vkmgw" Dec 03 22:09:54 crc kubenswrapper[4715]: I1203 22:09:54.634423 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:09:54 crc kubenswrapper[4715]: E1203 22:09:54.634778 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:09:59 crc kubenswrapper[4715]: I1203 22:09:59.786329 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-4d8n9" Dec 03 22:10:06 crc kubenswrapper[4715]: I1203 22:10:06.634906 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:10:06 crc kubenswrapper[4715]: E1203 22:10:06.636047 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:10:14 crc kubenswrapper[4715]: I1203 22:10:14.863830 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-sppb5" podUID="fa7301e8-1cff-40a6-a83c-d8e2db119bce" containerName="console" containerID="cri-o://5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac" gracePeriod=15 Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.864795 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sppb5_fa7301e8-1cff-40a6-a83c-d8e2db119bce/console/0.log" Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.865062 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sppb5" Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.901011 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-sppb5_fa7301e8-1cff-40a6-a83c-d8e2db119bce/console/0.log" Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.901059 4715 generic.go:334] "Generic (PLEG): container finished" podID="fa7301e8-1cff-40a6-a83c-d8e2db119bce" containerID="5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac" exitCode=2 Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.901084 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sppb5" event={"ID":"fa7301e8-1cff-40a6-a83c-d8e2db119bce","Type":"ContainerDied","Data":"5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac"} Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.901105 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-sppb5" event={"ID":"fa7301e8-1cff-40a6-a83c-d8e2db119bce","Type":"ContainerDied","Data":"4a5017defc8e37f5fe04667b690591659cd111f229e58b18bb90448a2de8e8c6"} Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.901117 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-sppb5" Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.901122 4715 scope.go:117] "RemoveContainer" containerID="5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac" Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.945436 4715 scope.go:117] "RemoveContainer" containerID="5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac" Dec 03 22:10:15 crc kubenswrapper[4715]: E1203 22:10:15.946158 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac\": container with ID starting with 5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac not found: ID does not exist" containerID="5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac" Dec 03 22:10:15 crc kubenswrapper[4715]: I1203 22:10:15.946195 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac"} err="failed to get container status \"5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac\": rpc error: code = NotFound desc = could not find container \"5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac\": container with ID starting with 5ef0966ea11e311a0478e8e69b42bf0eecc065ea7e4511b2eb923a30c7251eac not found: ID does not exist" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.060627 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-service-ca\") pod \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.060680 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-oauth-serving-cert\") pod \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.060709 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-serving-cert\") pod \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.060767 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-config\") pod \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.060841 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7ppj\" (UniqueName: \"kubernetes.io/projected/fa7301e8-1cff-40a6-a83c-d8e2db119bce-kube-api-access-n7ppj\") pod \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.060860 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-trusted-ca-bundle\") pod \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.060878 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-oauth-config\") pod \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\" (UID: \"fa7301e8-1cff-40a6-a83c-d8e2db119bce\") " Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.061713 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-service-ca" (OuterVolumeSpecName: "service-ca") pod "fa7301e8-1cff-40a6-a83c-d8e2db119bce" (UID: "fa7301e8-1cff-40a6-a83c-d8e2db119bce"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.061762 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "fa7301e8-1cff-40a6-a83c-d8e2db119bce" (UID: "fa7301e8-1cff-40a6-a83c-d8e2db119bce"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.061773 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "fa7301e8-1cff-40a6-a83c-d8e2db119bce" (UID: "fa7301e8-1cff-40a6-a83c-d8e2db119bce"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.061855 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-config" (OuterVolumeSpecName: "console-config") pod "fa7301e8-1cff-40a6-a83c-d8e2db119bce" (UID: "fa7301e8-1cff-40a6-a83c-d8e2db119bce"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.067339 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "fa7301e8-1cff-40a6-a83c-d8e2db119bce" (UID: "fa7301e8-1cff-40a6-a83c-d8e2db119bce"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.067858 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "fa7301e8-1cff-40a6-a83c-d8e2db119bce" (UID: "fa7301e8-1cff-40a6-a83c-d8e2db119bce"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.069145 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa7301e8-1cff-40a6-a83c-d8e2db119bce-kube-api-access-n7ppj" (OuterVolumeSpecName: "kube-api-access-n7ppj") pod "fa7301e8-1cff-40a6-a83c-d8e2db119bce" (UID: "fa7301e8-1cff-40a6-a83c-d8e2db119bce"). InnerVolumeSpecName "kube-api-access-n7ppj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.161795 4715 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.161831 4715 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.161841 4715 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.161850 4715 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.161859 4715 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.161868 4715 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/fa7301e8-1cff-40a6-a83c-d8e2db119bce-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.161876 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7ppj\" (UniqueName: \"kubernetes.io/projected/fa7301e8-1cff-40a6-a83c-d8e2db119bce-kube-api-access-n7ppj\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.234532 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-sppb5"] Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.243991 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-sppb5"] Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.393051 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h"] Dec 03 22:10:16 crc kubenswrapper[4715]: E1203 22:10:16.393328 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="registry-server" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.393343 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="registry-server" Dec 03 22:10:16 crc kubenswrapper[4715]: E1203 22:10:16.393360 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="extract-utilities" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.393368 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="extract-utilities" Dec 03 22:10:16 crc kubenswrapper[4715]: E1203 22:10:16.393377 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="extract-content" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.393388 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="extract-content" Dec 03 22:10:16 crc kubenswrapper[4715]: E1203 22:10:16.393397 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa7301e8-1cff-40a6-a83c-d8e2db119bce" containerName="console" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.393404 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa7301e8-1cff-40a6-a83c-d8e2db119bce" containerName="console" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.393558 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cedcea1-389c-4a34-bebb-e192f37d92a3" containerName="registry-server" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.393583 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa7301e8-1cff-40a6-a83c-d8e2db119bce" containerName="console" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.394826 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.401384 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.417829 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h"] Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.568207 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.568306 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jrlm\" (UniqueName: \"kubernetes.io/projected/680d52ba-8bcc-496c-9a08-c9b118c8cf76-kube-api-access-5jrlm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.568533 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.669321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.669454 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.669909 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jrlm\" (UniqueName: \"kubernetes.io/projected/680d52ba-8bcc-496c-9a08-c9b118c8cf76-kube-api-access-5jrlm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.670247 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.670336 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.690402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jrlm\" (UniqueName: \"kubernetes.io/projected/680d52ba-8bcc-496c-9a08-c9b118c8cf76-kube-api-access-5jrlm\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:16 crc kubenswrapper[4715]: I1203 22:10:16.712523 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:17 crc kubenswrapper[4715]: I1203 22:10:17.215640 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h"] Dec 03 22:10:17 crc kubenswrapper[4715]: W1203 22:10:17.220829 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod680d52ba_8bcc_496c_9a08_c9b118c8cf76.slice/crio-3e9b97365f48092eaf4af04a542319d02ea0b00bce0e920f71f6a956600b54c0 WatchSource:0}: Error finding container 3e9b97365f48092eaf4af04a542319d02ea0b00bce0e920f71f6a956600b54c0: Status 404 returned error can't find the container with id 3e9b97365f48092eaf4af04a542319d02ea0b00bce0e920f71f6a956600b54c0 Dec 03 22:10:17 crc kubenswrapper[4715]: I1203 22:10:17.647367 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa7301e8-1cff-40a6-a83c-d8e2db119bce" path="/var/lib/kubelet/pods/fa7301e8-1cff-40a6-a83c-d8e2db119bce/volumes" Dec 03 22:10:17 crc kubenswrapper[4715]: I1203 22:10:17.922854 4715 generic.go:334] "Generic (PLEG): container finished" podID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerID="ef31226fe5e7e2a72ba728a1564400267f5f5d7698866c061531fe4c9034128b" exitCode=0 Dec 03 22:10:17 crc kubenswrapper[4715]: I1203 22:10:17.922913 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" event={"ID":"680d52ba-8bcc-496c-9a08-c9b118c8cf76","Type":"ContainerDied","Data":"ef31226fe5e7e2a72ba728a1564400267f5f5d7698866c061531fe4c9034128b"} Dec 03 22:10:17 crc kubenswrapper[4715]: I1203 22:10:17.922950 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" event={"ID":"680d52ba-8bcc-496c-9a08-c9b118c8cf76","Type":"ContainerStarted","Data":"3e9b97365f48092eaf4af04a542319d02ea0b00bce0e920f71f6a956600b54c0"} Dec 03 22:10:19 crc kubenswrapper[4715]: I1203 22:10:19.939476 4715 generic.go:334] "Generic (PLEG): container finished" podID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerID="588b4a3117d2ada4ba470a4b7c1e4a2fadd2c0a3f39efd4bf3678a246969258f" exitCode=0 Dec 03 22:10:19 crc kubenswrapper[4715]: I1203 22:10:19.939703 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" event={"ID":"680d52ba-8bcc-496c-9a08-c9b118c8cf76","Type":"ContainerDied","Data":"588b4a3117d2ada4ba470a4b7c1e4a2fadd2c0a3f39efd4bf3678a246969258f"} Dec 03 22:10:20 crc kubenswrapper[4715]: I1203 22:10:20.634254 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:10:20 crc kubenswrapper[4715]: E1203 22:10:20.634480 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:10:20 crc kubenswrapper[4715]: I1203 22:10:20.953362 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" event={"ID":"680d52ba-8bcc-496c-9a08-c9b118c8cf76","Type":"ContainerStarted","Data":"228558a724ec65b1d8c373060c07bf9ff2235482222a4fa5c4f1da97bfa42064"} Dec 03 22:10:20 crc kubenswrapper[4715]: I1203 22:10:20.978934 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" podStartSLOduration=3.689006526 podStartE2EDuration="4.97891359s" podCreationTimestamp="2025-12-03 22:10:16 +0000 UTC" firstStartedPulling="2025-12-03 22:10:17.925281868 +0000 UTC m=+1714.667992503" lastFinishedPulling="2025-12-03 22:10:19.215188932 +0000 UTC m=+1715.957899567" observedRunningTime="2025-12-03 22:10:20.976805145 +0000 UTC m=+1717.719515770" watchObservedRunningTime="2025-12-03 22:10:20.97891359 +0000 UTC m=+1717.721624185" Dec 03 22:10:21 crc kubenswrapper[4715]: I1203 22:10:21.962730 4715 generic.go:334] "Generic (PLEG): container finished" podID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerID="228558a724ec65b1d8c373060c07bf9ff2235482222a4fa5c4f1da97bfa42064" exitCode=0 Dec 03 22:10:21 crc kubenswrapper[4715]: I1203 22:10:21.962860 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" event={"ID":"680d52ba-8bcc-496c-9a08-c9b118c8cf76","Type":"ContainerDied","Data":"228558a724ec65b1d8c373060c07bf9ff2235482222a4fa5c4f1da97bfa42064"} Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.246478 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.377872 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-util\") pod \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.377965 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-bundle\") pod \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.378070 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jrlm\" (UniqueName: \"kubernetes.io/projected/680d52ba-8bcc-496c-9a08-c9b118c8cf76-kube-api-access-5jrlm\") pod \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\" (UID: \"680d52ba-8bcc-496c-9a08-c9b118c8cf76\") " Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.379846 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-bundle" (OuterVolumeSpecName: "bundle") pod "680d52ba-8bcc-496c-9a08-c9b118c8cf76" (UID: "680d52ba-8bcc-496c-9a08-c9b118c8cf76"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.385174 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/680d52ba-8bcc-496c-9a08-c9b118c8cf76-kube-api-access-5jrlm" (OuterVolumeSpecName: "kube-api-access-5jrlm") pod "680d52ba-8bcc-496c-9a08-c9b118c8cf76" (UID: "680d52ba-8bcc-496c-9a08-c9b118c8cf76"). InnerVolumeSpecName "kube-api-access-5jrlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.392375 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-util" (OuterVolumeSpecName: "util") pod "680d52ba-8bcc-496c-9a08-c9b118c8cf76" (UID: "680d52ba-8bcc-496c-9a08-c9b118c8cf76"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.480710 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jrlm\" (UniqueName: \"kubernetes.io/projected/680d52ba-8bcc-496c-9a08-c9b118c8cf76-kube-api-access-5jrlm\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.480751 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-util\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.480764 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/680d52ba-8bcc-496c-9a08-c9b118c8cf76-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.982148 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" event={"ID":"680d52ba-8bcc-496c-9a08-c9b118c8cf76","Type":"ContainerDied","Data":"3e9b97365f48092eaf4af04a542319d02ea0b00bce0e920f71f6a956600b54c0"} Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.982568 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e9b97365f48092eaf4af04a542319d02ea0b00bce0e920f71f6a956600b54c0" Dec 03 22:10:23 crc kubenswrapper[4715]: I1203 22:10:23.982256 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.446487 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw"] Dec 03 22:10:31 crc kubenswrapper[4715]: E1203 22:10:31.447237 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerName="extract" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.447254 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerName="extract" Dec 03 22:10:31 crc kubenswrapper[4715]: E1203 22:10:31.447277 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerName="util" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.447286 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerName="util" Dec 03 22:10:31 crc kubenswrapper[4715]: E1203 22:10:31.447297 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerName="pull" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.447305 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerName="pull" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.447433 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="680d52ba-8bcc-496c-9a08-c9b118c8cf76" containerName="extract" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.447962 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.450124 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.450693 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.450796 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.451052 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.451235 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-sxx9b" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.466592 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw"] Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.590980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm6q5\" (UniqueName: \"kubernetes.io/projected/3483831a-11e2-4881-8447-fc127b2ec983-kube-api-access-fm6q5\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.591037 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3483831a-11e2-4881-8447-fc127b2ec983-apiservice-cert\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.591068 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3483831a-11e2-4881-8447-fc127b2ec983-webhook-cert\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.692130 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm6q5\" (UniqueName: \"kubernetes.io/projected/3483831a-11e2-4881-8447-fc127b2ec983-kube-api-access-fm6q5\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.692175 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3483831a-11e2-4881-8447-fc127b2ec983-apiservice-cert\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.692194 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3483831a-11e2-4881-8447-fc127b2ec983-webhook-cert\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.698132 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3483831a-11e2-4881-8447-fc127b2ec983-webhook-cert\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.698181 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3483831a-11e2-4881-8447-fc127b2ec983-apiservice-cert\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.710608 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm6q5\" (UniqueName: \"kubernetes.io/projected/3483831a-11e2-4881-8447-fc127b2ec983-kube-api-access-fm6q5\") pod \"metallb-operator-controller-manager-b57df49df-j4jvw\" (UID: \"3483831a-11e2-4881-8447-fc127b2ec983\") " pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.764200 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.775981 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn"] Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.777149 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.778724 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-wv2cl" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.779045 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.779661 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.846487 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn"] Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.894948 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drthb\" (UniqueName: \"kubernetes.io/projected/0b23c7d6-56da-4a46-ab0e-e98665166baa-kube-api-access-drthb\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.895021 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b23c7d6-56da-4a46-ab0e-e98665166baa-apiservice-cert\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.895045 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b23c7d6-56da-4a46-ab0e-e98665166baa-webhook-cert\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.996237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b23c7d6-56da-4a46-ab0e-e98665166baa-apiservice-cert\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.996297 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b23c7d6-56da-4a46-ab0e-e98665166baa-webhook-cert\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:31 crc kubenswrapper[4715]: I1203 22:10:31.996416 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drthb\" (UniqueName: \"kubernetes.io/projected/0b23c7d6-56da-4a46-ab0e-e98665166baa-kube-api-access-drthb\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:32 crc kubenswrapper[4715]: I1203 22:10:32.002159 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0b23c7d6-56da-4a46-ab0e-e98665166baa-apiservice-cert\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:32 crc kubenswrapper[4715]: I1203 22:10:32.005966 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0b23c7d6-56da-4a46-ab0e-e98665166baa-webhook-cert\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:32 crc kubenswrapper[4715]: I1203 22:10:32.013088 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drthb\" (UniqueName: \"kubernetes.io/projected/0b23c7d6-56da-4a46-ab0e-e98665166baa-kube-api-access-drthb\") pod \"metallb-operator-webhook-server-7f9b456564-mq9gn\" (UID: \"0b23c7d6-56da-4a46-ab0e-e98665166baa\") " pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:32 crc kubenswrapper[4715]: I1203 22:10:32.116774 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw"] Dec 03 22:10:32 crc kubenswrapper[4715]: I1203 22:10:32.132740 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:32 crc kubenswrapper[4715]: I1203 22:10:32.588019 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn"] Dec 03 22:10:32 crc kubenswrapper[4715]: W1203 22:10:32.596573 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b23c7d6_56da_4a46_ab0e_e98665166baa.slice/crio-199c210009d231c5519779cc2266ed1ee2d20963e22373475f588c604299ca8b WatchSource:0}: Error finding container 199c210009d231c5519779cc2266ed1ee2d20963e22373475f588c604299ca8b: Status 404 returned error can't find the container with id 199c210009d231c5519779cc2266ed1ee2d20963e22373475f588c604299ca8b Dec 03 22:10:33 crc kubenswrapper[4715]: I1203 22:10:33.035258 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" event={"ID":"0b23c7d6-56da-4a46-ab0e-e98665166baa","Type":"ContainerStarted","Data":"199c210009d231c5519779cc2266ed1ee2d20963e22373475f588c604299ca8b"} Dec 03 22:10:33 crc kubenswrapper[4715]: I1203 22:10:33.036242 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" event={"ID":"3483831a-11e2-4881-8447-fc127b2ec983","Type":"ContainerStarted","Data":"5b6173107b3556fddd68cad1ca06f28bc47428f7c1d09ac516e86ebbc9243b02"} Dec 03 22:10:34 crc kubenswrapper[4715]: I1203 22:10:34.635022 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:10:34 crc kubenswrapper[4715]: E1203 22:10:34.635823 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:10:41 crc kubenswrapper[4715]: I1203 22:10:41.101777 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" event={"ID":"3483831a-11e2-4881-8447-fc127b2ec983","Type":"ContainerStarted","Data":"141f631b2a38095fb46cedb2f64f4d99698d6a9d2bb4ca310360b1606f8e59ff"} Dec 03 22:10:41 crc kubenswrapper[4715]: I1203 22:10:41.102400 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:10:41 crc kubenswrapper[4715]: I1203 22:10:41.103561 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" event={"ID":"0b23c7d6-56da-4a46-ab0e-e98665166baa","Type":"ContainerStarted","Data":"e070408ead54325935cb94c4f0b78855aa205d328fec4ff7195899091c682aa3"} Dec 03 22:10:41 crc kubenswrapper[4715]: I1203 22:10:41.103693 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:10:41 crc kubenswrapper[4715]: I1203 22:10:41.156086 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" podStartSLOduration=2.563939841 podStartE2EDuration="10.156070285s" podCreationTimestamp="2025-12-03 22:10:31 +0000 UTC" firstStartedPulling="2025-12-03 22:10:32.601202196 +0000 UTC m=+1729.343912791" lastFinishedPulling="2025-12-03 22:10:40.19333264 +0000 UTC m=+1736.936043235" observedRunningTime="2025-12-03 22:10:41.154034821 +0000 UTC m=+1737.896745416" watchObservedRunningTime="2025-12-03 22:10:41.156070285 +0000 UTC m=+1737.898780880" Dec 03 22:10:41 crc kubenswrapper[4715]: I1203 22:10:41.156927 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" podStartSLOduration=2.116282475 podStartE2EDuration="10.156922007s" podCreationTimestamp="2025-12-03 22:10:31 +0000 UTC" firstStartedPulling="2025-12-03 22:10:32.132992979 +0000 UTC m=+1728.875703574" lastFinishedPulling="2025-12-03 22:10:40.173622671 +0000 UTC m=+1736.916343106" observedRunningTime="2025-12-03 22:10:41.126754903 +0000 UTC m=+1737.869465498" watchObservedRunningTime="2025-12-03 22:10:41.156922007 +0000 UTC m=+1737.899632602" Dec 03 22:10:49 crc kubenswrapper[4715]: I1203 22:10:49.634480 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:10:49 crc kubenswrapper[4715]: E1203 22:10:49.640981 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:10:52 crc kubenswrapper[4715]: I1203 22:10:52.137985 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7f9b456564-mq9gn" Dec 03 22:11:04 crc kubenswrapper[4715]: I1203 22:11:04.635298 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:11:04 crc kubenswrapper[4715]: E1203 22:11:04.636237 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:11:11 crc kubenswrapper[4715]: I1203 22:11:11.767399 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-b57df49df-j4jvw" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.729794 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-mvgh5"] Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.732822 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.736102 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.736528 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-g889g" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.749468 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9"] Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.750687 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.751355 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.767410 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.776791 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9"] Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.897142 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-rjg8s"] Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.898268 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-rjg8s" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.902628 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.904055 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-lrrj9" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.904224 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.905336 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909221 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-conf\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff247a17-7b16-4a5e-adcf-48f9122130eb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-94tb9\" (UID: \"ff247a17-7b16-4a5e-adcf-48f9122130eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-reloader\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909294 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-startup\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909321 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-sockets\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909342 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tzfj\" (UniqueName: \"kubernetes.io/projected/ff247a17-7b16-4a5e-adcf-48f9122130eb-kube-api-access-8tzfj\") pod \"frr-k8s-webhook-server-7fcb986d4-94tb9\" (UID: \"ff247a17-7b16-4a5e-adcf-48f9122130eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909368 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae01887b-8d01-4eaf-8214-3436b7f7af87-metrics-certs\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909386 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gxxk\" (UniqueName: \"kubernetes.io/projected/ae01887b-8d01-4eaf-8214-3436b7f7af87-kube-api-access-5gxxk\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.909435 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-metrics\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.916874 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-22qzs"] Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.917722 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.924323 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 22:11:12 crc kubenswrapper[4715]: I1203 22:11:12.932260 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-22qzs"] Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.010851 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-reloader\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.010902 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-startup\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.010937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-metrics-certs\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.010960 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-sockets\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.010990 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfnks\" (UniqueName: \"kubernetes.io/projected/b396984d-e5db-4417-ab0b-c2846cc8c95a-kube-api-access-cfnks\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011025 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tzfj\" (UniqueName: \"kubernetes.io/projected/ff247a17-7b16-4a5e-adcf-48f9122130eb-kube-api-access-8tzfj\") pod \"frr-k8s-webhook-server-7fcb986d4-94tb9\" (UID: \"ff247a17-7b16-4a5e-adcf-48f9122130eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011046 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b396984d-e5db-4417-ab0b-c2846cc8c95a-metallb-excludel2\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011068 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae01887b-8d01-4eaf-8214-3436b7f7af87-metrics-certs\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011084 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gxxk\" (UniqueName: \"kubernetes.io/projected/ae01887b-8d01-4eaf-8214-3436b7f7af87-kube-api-access-5gxxk\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011121 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-metrics-certs\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011143 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011161 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wfvs\" (UniqueName: \"kubernetes.io/projected/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-kube-api-access-4wfvs\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011190 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-metrics\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011205 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-conf\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011219 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff247a17-7b16-4a5e-adcf-48f9122130eb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-94tb9\" (UID: \"ff247a17-7b16-4a5e-adcf-48f9122130eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011235 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-cert\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.011654 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-reloader\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.012350 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-startup\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.012481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-sockets\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.012569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-metrics\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.012752 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/ae01887b-8d01-4eaf-8214-3436b7f7af87-frr-conf\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.018838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae01887b-8d01-4eaf-8214-3436b7f7af87-metrics-certs\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.021172 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ff247a17-7b16-4a5e-adcf-48f9122130eb-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-94tb9\" (UID: \"ff247a17-7b16-4a5e-adcf-48f9122130eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.033042 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gxxk\" (UniqueName: \"kubernetes.io/projected/ae01887b-8d01-4eaf-8214-3436b7f7af87-kube-api-access-5gxxk\") pod \"frr-k8s-mvgh5\" (UID: \"ae01887b-8d01-4eaf-8214-3436b7f7af87\") " pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.036786 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tzfj\" (UniqueName: \"kubernetes.io/projected/ff247a17-7b16-4a5e-adcf-48f9122130eb-kube-api-access-8tzfj\") pod \"frr-k8s-webhook-server-7fcb986d4-94tb9\" (UID: \"ff247a17-7b16-4a5e-adcf-48f9122130eb\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.050969 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.063208 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.112190 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-cert\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.112245 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-metrics-certs\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.112267 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfnks\" (UniqueName: \"kubernetes.io/projected/b396984d-e5db-4417-ab0b-c2846cc8c95a-kube-api-access-cfnks\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.112294 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b396984d-e5db-4417-ab0b-c2846cc8c95a-metallb-excludel2\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.112331 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-metrics-certs\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.112354 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.112371 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wfvs\" (UniqueName: \"kubernetes.io/projected/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-kube-api-access-4wfvs\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: E1203 22:11:13.113562 4715 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 03 22:11:13 crc kubenswrapper[4715]: E1203 22:11:13.113600 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 22:11:13 crc kubenswrapper[4715]: E1203 22:11:13.113640 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-metrics-certs podName:a7a0c0e5-fb39-479e-8731-f81bc225c7fb nodeName:}" failed. No retries permitted until 2025-12-03 22:11:13.613621001 +0000 UTC m=+1770.356331596 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-metrics-certs") pod "controller-f8648f98b-22qzs" (UID: "a7a0c0e5-fb39-479e-8731-f81bc225c7fb") : secret "controller-certs-secret" not found Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.113664 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/b396984d-e5db-4417-ab0b-c2846cc8c95a-metallb-excludel2\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: E1203 22:11:13.113692 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist podName:b396984d-e5db-4417-ab0b-c2846cc8c95a nodeName:}" failed. No retries permitted until 2025-12-03 22:11:13.613668402 +0000 UTC m=+1770.356379037 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist") pod "speaker-rjg8s" (UID: "b396984d-e5db-4417-ab0b-c2846cc8c95a") : secret "metallb-memberlist" not found Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.117321 4715 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.117494 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-metrics-certs\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.129069 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-cert\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.132873 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfnks\" (UniqueName: \"kubernetes.io/projected/b396984d-e5db-4417-ab0b-c2846cc8c95a-kube-api-access-cfnks\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.134318 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wfvs\" (UniqueName: \"kubernetes.io/projected/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-kube-api-access-4wfvs\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.373423 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9"] Dec 03 22:11:13 crc kubenswrapper[4715]: W1203 22:11:13.378336 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff247a17_7b16_4a5e_adcf_48f9122130eb.slice/crio-d0c1775a9cd14b41adee9adb5c3349ddc7771af02f86b47f0558af0754719137 WatchSource:0}: Error finding container d0c1775a9cd14b41adee9adb5c3349ddc7771af02f86b47f0558af0754719137: Status 404 returned error can't find the container with id d0c1775a9cd14b41adee9adb5c3349ddc7771af02f86b47f0558af0754719137 Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.620742 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-metrics-certs\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.620834 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:13 crc kubenswrapper[4715]: E1203 22:11:13.621126 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 22:11:13 crc kubenswrapper[4715]: E1203 22:11:13.621230 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist podName:b396984d-e5db-4417-ab0b-c2846cc8c95a nodeName:}" failed. No retries permitted until 2025-12-03 22:11:14.621201444 +0000 UTC m=+1771.363912079 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist") pod "speaker-rjg8s" (UID: "b396984d-e5db-4417-ab0b-c2846cc8c95a") : secret "metallb-memberlist" not found Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.626952 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a7a0c0e5-fb39-479e-8731-f81bc225c7fb-metrics-certs\") pod \"controller-f8648f98b-22qzs\" (UID: \"a7a0c0e5-fb39-479e-8731-f81bc225c7fb\") " pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:13 crc kubenswrapper[4715]: I1203 22:11:13.833131 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:14 crc kubenswrapper[4715]: I1203 22:11:14.279799 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-22qzs"] Dec 03 22:11:14 crc kubenswrapper[4715]: I1203 22:11:14.335383 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-22qzs" event={"ID":"a7a0c0e5-fb39-479e-8731-f81bc225c7fb","Type":"ContainerStarted","Data":"fe6da8ba59736ae93dd97de9d7a4626b6e2b0fa7d7a69ba1a4174b41e5c85380"} Dec 03 22:11:14 crc kubenswrapper[4715]: I1203 22:11:14.336383 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerStarted","Data":"df09b92edae16a2c0e27d0c4f54f3862bb6df2e068089e24da28cf733254e16c"} Dec 03 22:11:14 crc kubenswrapper[4715]: I1203 22:11:14.337350 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" event={"ID":"ff247a17-7b16-4a5e-adcf-48f9122130eb","Type":"ContainerStarted","Data":"d0c1775a9cd14b41adee9adb5c3349ddc7771af02f86b47f0558af0754719137"} Dec 03 22:11:14 crc kubenswrapper[4715]: I1203 22:11:14.636768 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:14 crc kubenswrapper[4715]: E1203 22:11:14.637052 4715 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 22:11:14 crc kubenswrapper[4715]: E1203 22:11:14.637128 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist podName:b396984d-e5db-4417-ab0b-c2846cc8c95a nodeName:}" failed. No retries permitted until 2025-12-03 22:11:16.637110849 +0000 UTC m=+1773.379821444 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist") pod "speaker-rjg8s" (UID: "b396984d-e5db-4417-ab0b-c2846cc8c95a") : secret "metallb-memberlist" not found Dec 03 22:11:15 crc kubenswrapper[4715]: I1203 22:11:15.358024 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-22qzs" event={"ID":"a7a0c0e5-fb39-479e-8731-f81bc225c7fb","Type":"ContainerStarted","Data":"6fbca1bd85155b2f10212501ddba808ded323f069d5cedb2ec528146d2b31662"} Dec 03 22:11:15 crc kubenswrapper[4715]: I1203 22:11:15.358070 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-22qzs" event={"ID":"a7a0c0e5-fb39-479e-8731-f81bc225c7fb","Type":"ContainerStarted","Data":"61f15f423a7122aa77dca0e8f77af791e6365093848aa94750937aba36554cdc"} Dec 03 22:11:15 crc kubenswrapper[4715]: I1203 22:11:15.358099 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:15 crc kubenswrapper[4715]: I1203 22:11:15.378741 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-22qzs" podStartSLOduration=3.378721778 podStartE2EDuration="3.378721778s" podCreationTimestamp="2025-12-03 22:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:11:15.378097852 +0000 UTC m=+1772.120808447" watchObservedRunningTime="2025-12-03 22:11:15.378721778 +0000 UTC m=+1772.121432373" Dec 03 22:11:16 crc kubenswrapper[4715]: I1203 22:11:16.665391 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:16 crc kubenswrapper[4715]: I1203 22:11:16.676459 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/b396984d-e5db-4417-ab0b-c2846cc8c95a-memberlist\") pod \"speaker-rjg8s\" (UID: \"b396984d-e5db-4417-ab0b-c2846cc8c95a\") " pod="metallb-system/speaker-rjg8s" Dec 03 22:11:16 crc kubenswrapper[4715]: I1203 22:11:16.812030 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-rjg8s" Dec 03 22:11:16 crc kubenswrapper[4715]: W1203 22:11:16.855355 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb396984d_e5db_4417_ab0b_c2846cc8c95a.slice/crio-1042004af6bdbe3e64b3bdf1c983a52ee2f0f8c5f78b595d51bfe937ca19257e WatchSource:0}: Error finding container 1042004af6bdbe3e64b3bdf1c983a52ee2f0f8c5f78b595d51bfe937ca19257e: Status 404 returned error can't find the container with id 1042004af6bdbe3e64b3bdf1c983a52ee2f0f8c5f78b595d51bfe937ca19257e Dec 03 22:11:17 crc kubenswrapper[4715]: I1203 22:11:17.371708 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rjg8s" event={"ID":"b396984d-e5db-4417-ab0b-c2846cc8c95a","Type":"ContainerStarted","Data":"23469d4e829c6a00b31f936cf2d6150cd12eab1500c92b7cd9e970daae2aeda1"} Dec 03 22:11:17 crc kubenswrapper[4715]: I1203 22:11:17.372199 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rjg8s" event={"ID":"b396984d-e5db-4417-ab0b-c2846cc8c95a","Type":"ContainerStarted","Data":"1042004af6bdbe3e64b3bdf1c983a52ee2f0f8c5f78b595d51bfe937ca19257e"} Dec 03 22:11:18 crc kubenswrapper[4715]: I1203 22:11:18.379351 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rjg8s" event={"ID":"b396984d-e5db-4417-ab0b-c2846cc8c95a","Type":"ContainerStarted","Data":"321ad9b3f382fd3b19b67c18ef43ab29e6db7719575383e1405899ca1165fe92"} Dec 03 22:11:18 crc kubenswrapper[4715]: I1203 22:11:18.379516 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-rjg8s" Dec 03 22:11:18 crc kubenswrapper[4715]: I1203 22:11:18.400997 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-rjg8s" podStartSLOduration=6.400978054 podStartE2EDuration="6.400978054s" podCreationTimestamp="2025-12-03 22:11:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:11:18.39705349 +0000 UTC m=+1775.139764085" watchObservedRunningTime="2025-12-03 22:11:18.400978054 +0000 UTC m=+1775.143688659" Dec 03 22:11:19 crc kubenswrapper[4715]: I1203 22:11:19.634261 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:11:19 crc kubenswrapper[4715]: E1203 22:11:19.634793 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:11:21 crc kubenswrapper[4715]: I1203 22:11:21.406268 4715 generic.go:334] "Generic (PLEG): container finished" podID="ae01887b-8d01-4eaf-8214-3436b7f7af87" containerID="50e314046efc7fee21cfe6961bfb75f1ad1e02b9ffa01958d775055610d88fe0" exitCode=0 Dec 03 22:11:21 crc kubenswrapper[4715]: I1203 22:11:21.406735 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerDied","Data":"50e314046efc7fee21cfe6961bfb75f1ad1e02b9ffa01958d775055610d88fe0"} Dec 03 22:11:21 crc kubenswrapper[4715]: I1203 22:11:21.410858 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" event={"ID":"ff247a17-7b16-4a5e-adcf-48f9122130eb","Type":"ContainerStarted","Data":"1572648a04f33a1fdebf467810247d58565ac94b1af96b2d4583b98d74271174"} Dec 03 22:11:21 crc kubenswrapper[4715]: I1203 22:11:21.411088 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:21 crc kubenswrapper[4715]: I1203 22:11:21.457455 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" podStartSLOduration=2.407562013 podStartE2EDuration="9.457429571s" podCreationTimestamp="2025-12-03 22:11:12 +0000 UTC" firstStartedPulling="2025-12-03 22:11:13.380182394 +0000 UTC m=+1770.122892989" lastFinishedPulling="2025-12-03 22:11:20.430049932 +0000 UTC m=+1777.172760547" observedRunningTime="2025-12-03 22:11:21.457130442 +0000 UTC m=+1778.199841077" watchObservedRunningTime="2025-12-03 22:11:21.457429571 +0000 UTC m=+1778.200140176" Dec 03 22:11:22 crc kubenswrapper[4715]: I1203 22:11:22.423221 4715 generic.go:334] "Generic (PLEG): container finished" podID="ae01887b-8d01-4eaf-8214-3436b7f7af87" containerID="7ab3e0b90e21ad4027a4568bb839d2bc5ef42139b2d045122a2d401498d207ae" exitCode=0 Dec 03 22:11:22 crc kubenswrapper[4715]: I1203 22:11:22.423387 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerDied","Data":"7ab3e0b90e21ad4027a4568bb839d2bc5ef42139b2d045122a2d401498d207ae"} Dec 03 22:11:24 crc kubenswrapper[4715]: I1203 22:11:24.442549 4715 generic.go:334] "Generic (PLEG): container finished" podID="ae01887b-8d01-4eaf-8214-3436b7f7af87" containerID="e8b68325643bc12018abe955cb1ebe583a0780a04c2b496f97fca8a9db4e97d8" exitCode=0 Dec 03 22:11:24 crc kubenswrapper[4715]: I1203 22:11:24.442605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerDied","Data":"e8b68325643bc12018abe955cb1ebe583a0780a04c2b496f97fca8a9db4e97d8"} Dec 03 22:11:25 crc kubenswrapper[4715]: I1203 22:11:25.463706 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerStarted","Data":"b00853cd514455c6798600d11f3f61b91c5a38b331fc2afdcba05e55521bbbad"} Dec 03 22:11:25 crc kubenswrapper[4715]: I1203 22:11:25.463806 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerStarted","Data":"798c7040039fadd6bd9a1aef6f314e335bc9c6229f902519b52400fc94d6aa2d"} Dec 03 22:11:25 crc kubenswrapper[4715]: I1203 22:11:25.463824 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerStarted","Data":"2ffd6f72111ca7935f4b3fd009a900f7bc44c2dc83cdac39f1349fb54c318059"} Dec 03 22:11:25 crc kubenswrapper[4715]: I1203 22:11:25.463842 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerStarted","Data":"cae55d0941dcda31d16788edd95bba3259d79a29faad0a4e7f7505b18e3b8c45"} Dec 03 22:11:26 crc kubenswrapper[4715]: I1203 22:11:26.477913 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerStarted","Data":"a3987e2bb214f5610368bc6a126fc27e20efe7adf9e530422d071bcfc3eba458"} Dec 03 22:11:26 crc kubenswrapper[4715]: I1203 22:11:26.478261 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mvgh5" event={"ID":"ae01887b-8d01-4eaf-8214-3436b7f7af87","Type":"ContainerStarted","Data":"626fd709c267e47b2f3687c19eead7db980edc45a9562830ecb6221d0bb44f06"} Dec 03 22:11:27 crc kubenswrapper[4715]: I1203 22:11:27.486637 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:27 crc kubenswrapper[4715]: I1203 22:11:27.519611 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-mvgh5" podStartSLOduration=9.287262237 podStartE2EDuration="15.519590376s" podCreationTimestamp="2025-12-03 22:11:12 +0000 UTC" firstStartedPulling="2025-12-03 22:11:14.203630849 +0000 UTC m=+1770.946341464" lastFinishedPulling="2025-12-03 22:11:20.435959008 +0000 UTC m=+1777.178669603" observedRunningTime="2025-12-03 22:11:27.516811102 +0000 UTC m=+1784.259521737" watchObservedRunningTime="2025-12-03 22:11:27.519590376 +0000 UTC m=+1784.262300971" Dec 03 22:11:28 crc kubenswrapper[4715]: I1203 22:11:28.051970 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:28 crc kubenswrapper[4715]: I1203 22:11:28.101297 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:30 crc kubenswrapper[4715]: I1203 22:11:30.634923 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:11:30 crc kubenswrapper[4715]: E1203 22:11:30.635796 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:11:33 crc kubenswrapper[4715]: I1203 22:11:33.075569 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-94tb9" Dec 03 22:11:33 crc kubenswrapper[4715]: I1203 22:11:33.838238 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-22qzs" Dec 03 22:11:36 crc kubenswrapper[4715]: I1203 22:11:36.817640 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-rjg8s" Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.271420 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-pmhqs"] Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.272256 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pmhqs" Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.276296 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-d98hl" Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.277312 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.277600 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.296518 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pmhqs"] Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.424311 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpx2r\" (UniqueName: \"kubernetes.io/projected/2a684fd3-b17c-42d2-b765-95fd7c56a32d-kube-api-access-cpx2r\") pod \"openstack-operator-index-pmhqs\" (UID: \"2a684fd3-b17c-42d2-b765-95fd7c56a32d\") " pod="openstack-operators/openstack-operator-index-pmhqs" Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.525657 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpx2r\" (UniqueName: \"kubernetes.io/projected/2a684fd3-b17c-42d2-b765-95fd7c56a32d-kube-api-access-cpx2r\") pod \"openstack-operator-index-pmhqs\" (UID: \"2a684fd3-b17c-42d2-b765-95fd7c56a32d\") " pod="openstack-operators/openstack-operator-index-pmhqs" Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.543402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpx2r\" (UniqueName: \"kubernetes.io/projected/2a684fd3-b17c-42d2-b765-95fd7c56a32d-kube-api-access-cpx2r\") pod \"openstack-operator-index-pmhqs\" (UID: \"2a684fd3-b17c-42d2-b765-95fd7c56a32d\") " pod="openstack-operators/openstack-operator-index-pmhqs" Dec 03 22:11:40 crc kubenswrapper[4715]: I1203 22:11:40.600617 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pmhqs" Dec 03 22:11:41 crc kubenswrapper[4715]: I1203 22:11:41.095573 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-pmhqs"] Dec 03 22:11:41 crc kubenswrapper[4715]: I1203 22:11:41.111913 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:11:41 crc kubenswrapper[4715]: I1203 22:11:41.595053 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pmhqs" event={"ID":"2a684fd3-b17c-42d2-b765-95fd7c56a32d","Type":"ContainerStarted","Data":"9a541201778d7e23b1b474ddd7ffd7bb42067e1a86f2434a39127d0478f837b5"} Dec 03 22:11:43 crc kubenswrapper[4715]: I1203 22:11:43.055666 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-mvgh5" Dec 03 22:11:43 crc kubenswrapper[4715]: I1203 22:11:43.659920 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-pmhqs"] Dec 03 22:11:44 crc kubenswrapper[4715]: I1203 22:11:44.252434 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-tpfhd"] Dec 03 22:11:44 crc kubenswrapper[4715]: I1203 22:11:44.253352 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:44 crc kubenswrapper[4715]: I1203 22:11:44.268591 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tpfhd"] Dec 03 22:11:44 crc kubenswrapper[4715]: I1203 22:11:44.377941 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krkws\" (UniqueName: \"kubernetes.io/projected/dc2ba455-421c-4b30-9a6c-93945d7d1c9f-kube-api-access-krkws\") pod \"openstack-operator-index-tpfhd\" (UID: \"dc2ba455-421c-4b30-9a6c-93945d7d1c9f\") " pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:44 crc kubenswrapper[4715]: I1203 22:11:44.482338 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krkws\" (UniqueName: \"kubernetes.io/projected/dc2ba455-421c-4b30-9a6c-93945d7d1c9f-kube-api-access-krkws\") pod \"openstack-operator-index-tpfhd\" (UID: \"dc2ba455-421c-4b30-9a6c-93945d7d1c9f\") " pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:44 crc kubenswrapper[4715]: I1203 22:11:44.512261 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krkws\" (UniqueName: \"kubernetes.io/projected/dc2ba455-421c-4b30-9a6c-93945d7d1c9f-kube-api-access-krkws\") pod \"openstack-operator-index-tpfhd\" (UID: \"dc2ba455-421c-4b30-9a6c-93945d7d1c9f\") " pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:44 crc kubenswrapper[4715]: I1203 22:11:44.568820 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:44 crc kubenswrapper[4715]: I1203 22:11:44.634220 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:11:44 crc kubenswrapper[4715]: E1203 22:11:44.634431 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:11:45 crc kubenswrapper[4715]: I1203 22:11:45.436561 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-tpfhd"] Dec 03 22:11:46 crc kubenswrapper[4715]: W1203 22:11:46.369032 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc2ba455_421c_4b30_9a6c_93945d7d1c9f.slice/crio-5a933240b9870fc8e0fc6ded1cadc3ecaf6ce2bcc78316ed0072698d45c6a919 WatchSource:0}: Error finding container 5a933240b9870fc8e0fc6ded1cadc3ecaf6ce2bcc78316ed0072698d45c6a919: Status 404 returned error can't find the container with id 5a933240b9870fc8e0fc6ded1cadc3ecaf6ce2bcc78316ed0072698d45c6a919 Dec 03 22:11:46 crc kubenswrapper[4715]: I1203 22:11:46.624679 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tpfhd" event={"ID":"dc2ba455-421c-4b30-9a6c-93945d7d1c9f","Type":"ContainerStarted","Data":"5a933240b9870fc8e0fc6ded1cadc3ecaf6ce2bcc78316ed0072698d45c6a919"} Dec 03 22:11:48 crc kubenswrapper[4715]: I1203 22:11:48.643581 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pmhqs" event={"ID":"2a684fd3-b17c-42d2-b765-95fd7c56a32d","Type":"ContainerStarted","Data":"e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae"} Dec 03 22:11:48 crc kubenswrapper[4715]: I1203 22:11:48.643633 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-pmhqs" podUID="2a684fd3-b17c-42d2-b765-95fd7c56a32d" containerName="registry-server" containerID="cri-o://e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae" gracePeriod=2 Dec 03 22:11:48 crc kubenswrapper[4715]: I1203 22:11:48.647670 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-tpfhd" event={"ID":"dc2ba455-421c-4b30-9a6c-93945d7d1c9f","Type":"ContainerStarted","Data":"d9722f5d755a544918ad0695e119ed8becc636aa54abf7b248361a7727fd8eba"} Dec 03 22:11:48 crc kubenswrapper[4715]: I1203 22:11:48.670866 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-pmhqs" podStartSLOduration=1.9518953749999999 podStartE2EDuration="8.670849765s" podCreationTimestamp="2025-12-03 22:11:40 +0000 UTC" firstStartedPulling="2025-12-03 22:11:41.111650408 +0000 UTC m=+1797.854361003" lastFinishedPulling="2025-12-03 22:11:47.830604798 +0000 UTC m=+1804.573315393" observedRunningTime="2025-12-03 22:11:48.665796052 +0000 UTC m=+1805.408506647" watchObservedRunningTime="2025-12-03 22:11:48.670849765 +0000 UTC m=+1805.413560360" Dec 03 22:11:48 crc kubenswrapper[4715]: I1203 22:11:48.679231 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-tpfhd" podStartSLOduration=3.222828176 podStartE2EDuration="4.679213256s" podCreationTimestamp="2025-12-03 22:11:44 +0000 UTC" firstStartedPulling="2025-12-03 22:11:46.375352387 +0000 UTC m=+1803.118062982" lastFinishedPulling="2025-12-03 22:11:47.831737447 +0000 UTC m=+1804.574448062" observedRunningTime="2025-12-03 22:11:48.678675771 +0000 UTC m=+1805.421386366" watchObservedRunningTime="2025-12-03 22:11:48.679213256 +0000 UTC m=+1805.421923851" Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.117565 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pmhqs" Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.259835 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpx2r\" (UniqueName: \"kubernetes.io/projected/2a684fd3-b17c-42d2-b765-95fd7c56a32d-kube-api-access-cpx2r\") pod \"2a684fd3-b17c-42d2-b765-95fd7c56a32d\" (UID: \"2a684fd3-b17c-42d2-b765-95fd7c56a32d\") " Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.268221 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a684fd3-b17c-42d2-b765-95fd7c56a32d-kube-api-access-cpx2r" (OuterVolumeSpecName: "kube-api-access-cpx2r") pod "2a684fd3-b17c-42d2-b765-95fd7c56a32d" (UID: "2a684fd3-b17c-42d2-b765-95fd7c56a32d"). InnerVolumeSpecName "kube-api-access-cpx2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.362762 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpx2r\" (UniqueName: \"kubernetes.io/projected/2a684fd3-b17c-42d2-b765-95fd7c56a32d-kube-api-access-cpx2r\") on node \"crc\" DevicePath \"\"" Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.655343 4715 generic.go:334] "Generic (PLEG): container finished" podID="2a684fd3-b17c-42d2-b765-95fd7c56a32d" containerID="e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae" exitCode=0 Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.656582 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pmhqs" event={"ID":"2a684fd3-b17c-42d2-b765-95fd7c56a32d","Type":"ContainerDied","Data":"e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae"} Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.656605 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-pmhqs" Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.656654 4715 scope.go:117] "RemoveContainer" containerID="e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae" Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.656641 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-pmhqs" event={"ID":"2a684fd3-b17c-42d2-b765-95fd7c56a32d","Type":"ContainerDied","Data":"9a541201778d7e23b1b474ddd7ffd7bb42067e1a86f2434a39127d0478f837b5"} Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.686253 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-pmhqs"] Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.687426 4715 scope.go:117] "RemoveContainer" containerID="e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae" Dec 03 22:11:49 crc kubenswrapper[4715]: E1203 22:11:49.687797 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae\": container with ID starting with e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae not found: ID does not exist" containerID="e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae" Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.687838 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae"} err="failed to get container status \"e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae\": rpc error: code = NotFound desc = could not find container \"e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae\": container with ID starting with e211f01fe4d41cdaf8a18062e16072855218426015932eccf7e910a99c6dc0ae not found: ID does not exist" Dec 03 22:11:49 crc kubenswrapper[4715]: I1203 22:11:49.695078 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-pmhqs"] Dec 03 22:11:51 crc kubenswrapper[4715]: I1203 22:11:51.645334 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a684fd3-b17c-42d2-b765-95fd7c56a32d" path="/var/lib/kubelet/pods/2a684fd3-b17c-42d2-b765-95fd7c56a32d/volumes" Dec 03 22:11:54 crc kubenswrapper[4715]: I1203 22:11:54.569734 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:54 crc kubenswrapper[4715]: I1203 22:11:54.570218 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:54 crc kubenswrapper[4715]: I1203 22:11:54.610631 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:54 crc kubenswrapper[4715]: I1203 22:11:54.722832 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-tpfhd" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.516956 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88"] Dec 03 22:11:56 crc kubenswrapper[4715]: E1203 22:11:56.518349 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a684fd3-b17c-42d2-b765-95fd7c56a32d" containerName="registry-server" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.518380 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a684fd3-b17c-42d2-b765-95fd7c56a32d" containerName="registry-server" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.518684 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a684fd3-b17c-42d2-b765-95fd7c56a32d" containerName="registry-server" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.520556 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.524657 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-bw7pw" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.539416 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88"] Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.604063 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.604147 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.604264 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86l65\" (UniqueName: \"kubernetes.io/projected/cc5a2a79-3856-406e-88ac-b6496ea27a64-kube-api-access-86l65\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.706593 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.706724 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.706770 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86l65\" (UniqueName: \"kubernetes.io/projected/cc5a2a79-3856-406e-88ac-b6496ea27a64-kube-api-access-86l65\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.707398 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-bundle\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.707429 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-util\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.740795 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86l65\" (UniqueName: \"kubernetes.io/projected/cc5a2a79-3856-406e-88ac-b6496ea27a64-kube-api-access-86l65\") pod \"464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:56 crc kubenswrapper[4715]: I1203 22:11:56.853311 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:11:57 crc kubenswrapper[4715]: I1203 22:11:57.309242 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88"] Dec 03 22:11:57 crc kubenswrapper[4715]: I1203 22:11:57.721333 4715 generic.go:334] "Generic (PLEG): container finished" podID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerID="703a09db2c87e7e177f47eae34447c3c683f5431792c3f4a2f749863e9ad26ff" exitCode=0 Dec 03 22:11:57 crc kubenswrapper[4715]: I1203 22:11:57.721415 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" event={"ID":"cc5a2a79-3856-406e-88ac-b6496ea27a64","Type":"ContainerDied","Data":"703a09db2c87e7e177f47eae34447c3c683f5431792c3f4a2f749863e9ad26ff"} Dec 03 22:11:57 crc kubenswrapper[4715]: I1203 22:11:57.721695 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" event={"ID":"cc5a2a79-3856-406e-88ac-b6496ea27a64","Type":"ContainerStarted","Data":"785b1df93be95ec71329561b771ae547a641f83cf6905520abfdf5dbe5b87189"} Dec 03 22:11:58 crc kubenswrapper[4715]: I1203 22:11:58.634612 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:11:58 crc kubenswrapper[4715]: E1203 22:11:58.635156 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:11:58 crc kubenswrapper[4715]: I1203 22:11:58.733133 4715 generic.go:334] "Generic (PLEG): container finished" podID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerID="1fab7cea54af26fcd24acebd89abbc19dbc395a4c65f79591b088efe6d591457" exitCode=0 Dec 03 22:11:58 crc kubenswrapper[4715]: I1203 22:11:58.733193 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" event={"ID":"cc5a2a79-3856-406e-88ac-b6496ea27a64","Type":"ContainerDied","Data":"1fab7cea54af26fcd24acebd89abbc19dbc395a4c65f79591b088efe6d591457"} Dec 03 22:11:59 crc kubenswrapper[4715]: I1203 22:11:59.743432 4715 generic.go:334] "Generic (PLEG): container finished" podID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerID="aadeaab05ab0dea35f6cbed3b82859b2a3fef20b8f24f72155a3a64dcf84a1c0" exitCode=0 Dec 03 22:11:59 crc kubenswrapper[4715]: I1203 22:11:59.743551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" event={"ID":"cc5a2a79-3856-406e-88ac-b6496ea27a64","Type":"ContainerDied","Data":"aadeaab05ab0dea35f6cbed3b82859b2a3fef20b8f24f72155a3a64dcf84a1c0"} Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.064684 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.183000 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-util\") pod \"cc5a2a79-3856-406e-88ac-b6496ea27a64\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.183133 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86l65\" (UniqueName: \"kubernetes.io/projected/cc5a2a79-3856-406e-88ac-b6496ea27a64-kube-api-access-86l65\") pod \"cc5a2a79-3856-406e-88ac-b6496ea27a64\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.183326 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-bundle\") pod \"cc5a2a79-3856-406e-88ac-b6496ea27a64\" (UID: \"cc5a2a79-3856-406e-88ac-b6496ea27a64\") " Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.184078 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-bundle" (OuterVolumeSpecName: "bundle") pod "cc5a2a79-3856-406e-88ac-b6496ea27a64" (UID: "cc5a2a79-3856-406e-88ac-b6496ea27a64"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.191067 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc5a2a79-3856-406e-88ac-b6496ea27a64-kube-api-access-86l65" (OuterVolumeSpecName: "kube-api-access-86l65") pod "cc5a2a79-3856-406e-88ac-b6496ea27a64" (UID: "cc5a2a79-3856-406e-88ac-b6496ea27a64"). InnerVolumeSpecName "kube-api-access-86l65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.213072 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-util" (OuterVolumeSpecName: "util") pod "cc5a2a79-3856-406e-88ac-b6496ea27a64" (UID: "cc5a2a79-3856-406e-88ac-b6496ea27a64"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.284659 4715 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.284712 4715 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc5a2a79-3856-406e-88ac-b6496ea27a64-util\") on node \"crc\" DevicePath \"\"" Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.284731 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86l65\" (UniqueName: \"kubernetes.io/projected/cc5a2a79-3856-406e-88ac-b6496ea27a64-kube-api-access-86l65\") on node \"crc\" DevicePath \"\"" Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.763298 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.763407 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88" event={"ID":"cc5a2a79-3856-406e-88ac-b6496ea27a64","Type":"ContainerDied","Data":"785b1df93be95ec71329561b771ae547a641f83cf6905520abfdf5dbe5b87189"} Dec 03 22:12:01 crc kubenswrapper[4715]: I1203 22:12:01.763487 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="785b1df93be95ec71329561b771ae547a641f83cf6905520abfdf5dbe5b87189" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.605672 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x"] Dec 03 22:12:08 crc kubenswrapper[4715]: E1203 22:12:08.606473 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerName="util" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.606489 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerName="util" Dec 03 22:12:08 crc kubenswrapper[4715]: E1203 22:12:08.606525 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerName="pull" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.606532 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerName="pull" Dec 03 22:12:08 crc kubenswrapper[4715]: E1203 22:12:08.606544 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerName="extract" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.606552 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerName="extract" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.606651 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc5a2a79-3856-406e-88ac-b6496ea27a64" containerName="extract" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.607020 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.610100 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-5lzqd" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.644777 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x"] Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.719886 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dgwh\" (UniqueName: \"kubernetes.io/projected/d4195e77-3226-4e2f-b29e-b6ddaaa3bef2-kube-api-access-6dgwh\") pod \"openstack-operator-controller-operator-85f8f75cbc-4ld8x\" (UID: \"d4195e77-3226-4e2f-b29e-b6ddaaa3bef2\") " pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.821205 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dgwh\" (UniqueName: \"kubernetes.io/projected/d4195e77-3226-4e2f-b29e-b6ddaaa3bef2-kube-api-access-6dgwh\") pod \"openstack-operator-controller-operator-85f8f75cbc-4ld8x\" (UID: \"d4195e77-3226-4e2f-b29e-b6ddaaa3bef2\") " pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.849984 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dgwh\" (UniqueName: \"kubernetes.io/projected/d4195e77-3226-4e2f-b29e-b6ddaaa3bef2-kube-api-access-6dgwh\") pod \"openstack-operator-controller-operator-85f8f75cbc-4ld8x\" (UID: \"d4195e77-3226-4e2f-b29e-b6ddaaa3bef2\") " pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" Dec 03 22:12:08 crc kubenswrapper[4715]: I1203 22:12:08.928153 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" Dec 03 22:12:09 crc kubenswrapper[4715]: I1203 22:12:09.253604 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x"] Dec 03 22:12:09 crc kubenswrapper[4715]: I1203 22:12:09.634868 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:12:09 crc kubenswrapper[4715]: E1203 22:12:09.635265 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:12:09 crc kubenswrapper[4715]: I1203 22:12:09.822101 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" event={"ID":"d4195e77-3226-4e2f-b29e-b6ddaaa3bef2","Type":"ContainerStarted","Data":"4930fb434d7ca5ec9d07a7fcee34464a8f3e883f8a6b6c7090856df92c9d49ba"} Dec 03 22:12:20 crc kubenswrapper[4715]: I1203 22:12:20.913189 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" event={"ID":"d4195e77-3226-4e2f-b29e-b6ddaaa3bef2","Type":"ContainerStarted","Data":"47ab6095e2a8c60a901bea5a8cb44c821296d7cf2c53f038096ef84f7b50f975"} Dec 03 22:12:20 crc kubenswrapper[4715]: I1203 22:12:20.915063 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" Dec 03 22:12:20 crc kubenswrapper[4715]: I1203 22:12:20.951369 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" podStartSLOduration=1.9091107269999998 podStartE2EDuration="12.951352189s" podCreationTimestamp="2025-12-03 22:12:08 +0000 UTC" firstStartedPulling="2025-12-03 22:12:09.245737699 +0000 UTC m=+1825.988448304" lastFinishedPulling="2025-12-03 22:12:20.287979171 +0000 UTC m=+1837.030689766" observedRunningTime="2025-12-03 22:12:20.939556618 +0000 UTC m=+1837.682267263" watchObservedRunningTime="2025-12-03 22:12:20.951352189 +0000 UTC m=+1837.694062784" Dec 03 22:12:21 crc kubenswrapper[4715]: I1203 22:12:21.634042 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:12:21 crc kubenswrapper[4715]: E1203 22:12:21.634546 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:12:28 crc kubenswrapper[4715]: I1203 22:12:28.932621 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-85f8f75cbc-4ld8x" Dec 03 22:12:32 crc kubenswrapper[4715]: I1203 22:12:32.634548 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:12:32 crc kubenswrapper[4715]: E1203 22:12:32.635304 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:12:47 crc kubenswrapper[4715]: I1203 22:12:47.634157 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.077846 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.078909 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.082894 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-6rjlr" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.086024 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.087187 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.092888 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-vdpjs" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.102986 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.107488 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.108832 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-hsc22" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.123288 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.127235 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.134366 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.135654 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.139591 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.140380 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-czhs6" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.157023 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.158285 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.163003 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-bbss8" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.172618 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.180307 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.210677 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.211278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlkps\" (UniqueName: \"kubernetes.io/projected/46e49f12-51e7-45ca-a0cb-c7f8a40adf12-kube-api-access-tlkps\") pod \"barbican-operator-controller-manager-7d9dfd778-5hzjb\" (UID: \"46e49f12-51e7-45ca-a0cb-c7f8a40adf12\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.211352 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdgh9\" (UniqueName: \"kubernetes.io/projected/a91ec3b7-705a-4500-be8e-e8c789a2c973-kube-api-access-wdgh9\") pod \"designate-operator-controller-manager-78b4bc895b-sctl9\" (UID: \"a91ec3b7-705a-4500-be8e-e8c789a2c973\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.211383 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-988s9\" (UniqueName: \"kubernetes.io/projected/9c8c4875-3433-4885-9afa-acbfea9d546b-kube-api-access-988s9\") pod \"cinder-operator-controller-manager-859b6ccc6-mnd86\" (UID: \"9c8c4875-3433-4885-9afa-acbfea9d546b\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.211691 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.216741 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-dv68k" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.228290 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.243194 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.244174 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.248750 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.248900 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4nkj7" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.254180 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.254965 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.258580 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-f8fmt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.267357 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.285671 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.296513 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.297984 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.310428 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.310707 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6f2p8" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.312396 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-988s9\" (UniqueName: \"kubernetes.io/projected/9c8c4875-3433-4885-9afa-acbfea9d546b-kube-api-access-988s9\") pod \"cinder-operator-controller-manager-859b6ccc6-mnd86\" (UID: \"9c8c4875-3433-4885-9afa-acbfea9d546b\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.312447 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sgcc\" (UniqueName: \"kubernetes.io/projected/b8cbfb12-17c2-46bf-baea-ee8ac16db4ee-kube-api-access-6sgcc\") pod \"glance-operator-controller-manager-77987cd8cd-d5phr\" (UID: \"b8cbfb12-17c2-46bf-baea-ee8ac16db4ee\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.312468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfrdp\" (UniqueName: \"kubernetes.io/projected/1cba7fbc-0a66-45b4-b20d-33c81d407b96-kube-api-access-gfrdp\") pod \"heat-operator-controller-manager-5f64f6f8bb-gc2vn\" (UID: \"1cba7fbc-0a66-45b4-b20d-33c81d407b96\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.312519 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlkps\" (UniqueName: \"kubernetes.io/projected/46e49f12-51e7-45ca-a0cb-c7f8a40adf12-kube-api-access-tlkps\") pod \"barbican-operator-controller-manager-7d9dfd778-5hzjb\" (UID: \"46e49f12-51e7-45ca-a0cb-c7f8a40adf12\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.312543 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8hxk\" (UniqueName: \"kubernetes.io/projected/8219810d-511e-4cc8-a018-49f425b339c0-kube-api-access-s8hxk\") pod \"horizon-operator-controller-manager-68c6d99b8f-tgkbx\" (UID: \"8219810d-511e-4cc8-a018-49f425b339c0\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.312580 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdgh9\" (UniqueName: \"kubernetes.io/projected/a91ec3b7-705a-4500-be8e-e8c789a2c973-kube-api-access-wdgh9\") pod \"designate-operator-controller-manager-78b4bc895b-sctl9\" (UID: \"a91ec3b7-705a-4500-be8e-e8c789a2c973\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.312903 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.318103 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-82rwl" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.324001 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.329132 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.347484 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.354608 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-988s9\" (UniqueName: \"kubernetes.io/projected/9c8c4875-3433-4885-9afa-acbfea9d546b-kube-api-access-988s9\") pod \"cinder-operator-controller-manager-859b6ccc6-mnd86\" (UID: \"9c8c4875-3433-4885-9afa-acbfea9d546b\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.366135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdgh9\" (UniqueName: \"kubernetes.io/projected/a91ec3b7-705a-4500-be8e-e8c789a2c973-kube-api-access-wdgh9\") pod \"designate-operator-controller-manager-78b4bc895b-sctl9\" (UID: \"a91ec3b7-705a-4500-be8e-e8c789a2c973\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.369287 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.375895 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlkps\" (UniqueName: \"kubernetes.io/projected/46e49f12-51e7-45ca-a0cb-c7f8a40adf12-kube-api-access-tlkps\") pod \"barbican-operator-controller-manager-7d9dfd778-5hzjb\" (UID: \"46e49f12-51e7-45ca-a0cb-c7f8a40adf12\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.381187 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-x88ln" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.398538 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.407714 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.414795 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.415842 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.416250 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96mv5\" (UniqueName: \"kubernetes.io/projected/9a900ae2-08bc-403b-a613-52e7870eeebf-kube-api-access-96mv5\") pod \"manila-operator-controller-manager-7c79b5df47-9gf6m\" (UID: \"9a900ae2-08bc-403b-a613-52e7870eeebf\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.416281 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mxtk\" (UniqueName: \"kubernetes.io/projected/305a708b-69fc-4263-a28f-015fe16402f7-kube-api-access-2mxtk\") pod \"keystone-operator-controller-manager-7765d96ddf-qbp4s\" (UID: \"305a708b-69fc-4263-a28f-015fe16402f7\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.416304 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sgcc\" (UniqueName: \"kubernetes.io/projected/b8cbfb12-17c2-46bf-baea-ee8ac16db4ee-kube-api-access-6sgcc\") pod \"glance-operator-controller-manager-77987cd8cd-d5phr\" (UID: \"b8cbfb12-17c2-46bf-baea-ee8ac16db4ee\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.416325 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfrdp\" (UniqueName: \"kubernetes.io/projected/1cba7fbc-0a66-45b4-b20d-33c81d407b96-kube-api-access-gfrdp\") pod \"heat-operator-controller-manager-5f64f6f8bb-gc2vn\" (UID: \"1cba7fbc-0a66-45b4-b20d-33c81d407b96\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.416345 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.416413 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8hxk\" (UniqueName: \"kubernetes.io/projected/8219810d-511e-4cc8-a018-49f425b339c0-kube-api-access-s8hxk\") pod \"horizon-operator-controller-manager-68c6d99b8f-tgkbx\" (UID: \"8219810d-511e-4cc8-a018-49f425b339c0\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.416430 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dfff\" (UniqueName: \"kubernetes.io/projected/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-kube-api-access-9dfff\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.416455 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qqlt\" (UniqueName: \"kubernetes.io/projected/95b90127-8325-43c8-977e-8b972b62db2a-kube-api-access-8qqlt\") pod \"ironic-operator-controller-manager-6c548fd776-9qbjh\" (UID: \"95b90127-8325-43c8-977e-8b972b62db2a\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.420346 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-kz2mg" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.423896 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.424282 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.425358 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.439918 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.445987 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-svhz2" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.448122 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.449262 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.450193 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.451712 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-hsq8s" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.453088 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfrdp\" (UniqueName: \"kubernetes.io/projected/1cba7fbc-0a66-45b4-b20d-33c81d407b96-kube-api-access-gfrdp\") pod \"heat-operator-controller-manager-5f64f6f8bb-gc2vn\" (UID: \"1cba7fbc-0a66-45b4-b20d-33c81d407b96\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.456562 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sgcc\" (UniqueName: \"kubernetes.io/projected/b8cbfb12-17c2-46bf-baea-ee8ac16db4ee-kube-api-access-6sgcc\") pod \"glance-operator-controller-manager-77987cd8cd-d5phr\" (UID: \"b8cbfb12-17c2-46bf-baea-ee8ac16db4ee\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.458300 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.462772 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.467134 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8hxk\" (UniqueName: \"kubernetes.io/projected/8219810d-511e-4cc8-a018-49f425b339c0-kube-api-access-s8hxk\") pod \"horizon-operator-controller-manager-68c6d99b8f-tgkbx\" (UID: \"8219810d-511e-4cc8-a018-49f425b339c0\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.467180 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.472694 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.473785 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.477498 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-pqd8g" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.477702 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.477849 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.481851 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.487249 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.487301 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.487420 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.488879 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-dp27w" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.500211 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.501575 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.503676 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-27bx8" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.509953 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.512441 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.517790 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96mv5\" (UniqueName: \"kubernetes.io/projected/9a900ae2-08bc-403b-a613-52e7870eeebf-kube-api-access-96mv5\") pod \"manila-operator-controller-manager-7c79b5df47-9gf6m\" (UID: \"9a900ae2-08bc-403b-a613-52e7870eeebf\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.517825 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mxtk\" (UniqueName: \"kubernetes.io/projected/305a708b-69fc-4263-a28f-015fe16402f7-kube-api-access-2mxtk\") pod \"keystone-operator-controller-manager-7765d96ddf-qbp4s\" (UID: \"305a708b-69fc-4263-a28f-015fe16402f7\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.517853 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68rmm\" (UniqueName: \"kubernetes.io/projected/3091b003-a466-4d34-8a26-d3e68a3639a9-kube-api-access-68rmm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-gv4jj\" (UID: \"3091b003-a466-4d34-8a26-d3e68a3639a9\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.517913 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlgv4\" (UniqueName: \"kubernetes.io/projected/c6d7646e-fb4a-4199-aadb-670000485477-kube-api-access-jlgv4\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hqgrk\" (UID: \"c6d7646e-fb4a-4199-aadb-670000485477\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.517933 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.517986 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dfff\" (UniqueName: \"kubernetes.io/projected/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-kube-api-access-9dfff\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.518005 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqk6d\" (UniqueName: \"kubernetes.io/projected/b8304681-abf0-4e5c-aa37-41453f6e7731-kube-api-access-cqk6d\") pod \"nova-operator-controller-manager-697bc559fc-hm2q2\" (UID: \"b8304681-abf0-4e5c-aa37-41453f6e7731\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.518022 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qqlt\" (UniqueName: \"kubernetes.io/projected/95b90127-8325-43c8-977e-8b972b62db2a-kube-api-access-8qqlt\") pod \"ironic-operator-controller-manager-6c548fd776-9qbjh\" (UID: \"95b90127-8325-43c8-977e-8b972b62db2a\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.518117 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-4rd52" Dec 03 22:12:48 crc kubenswrapper[4715]: E1203 22:12:48.518121 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:48 crc kubenswrapper[4715]: E1203 22:12:48.518328 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert podName:6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef nodeName:}" failed. No retries permitted until 2025-12-03 22:12:49.018305968 +0000 UTC m=+1865.761016633 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert") pod "infra-operator-controller-manager-57548d458d-kwnlp" (UID: "6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef") : secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.524547 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.537647 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.538827 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96mv5\" (UniqueName: \"kubernetes.io/projected/9a900ae2-08bc-403b-a613-52e7870eeebf-kube-api-access-96mv5\") pod \"manila-operator-controller-manager-7c79b5df47-9gf6m\" (UID: \"9a900ae2-08bc-403b-a613-52e7870eeebf\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.543005 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qqlt\" (UniqueName: \"kubernetes.io/projected/95b90127-8325-43c8-977e-8b972b62db2a-kube-api-access-8qqlt\") pod \"ironic-operator-controller-manager-6c548fd776-9qbjh\" (UID: \"95b90127-8325-43c8-977e-8b972b62db2a\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.552156 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.566421 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.567570 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.572948 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-g9g7s" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.575457 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.582938 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.587642 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dfff\" (UniqueName: \"kubernetes.io/projected/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-kube-api-access-9dfff\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.590134 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mxtk\" (UniqueName: \"kubernetes.io/projected/305a708b-69fc-4263-a28f-015fe16402f7-kube-api-access-2mxtk\") pod \"keystone-operator-controller-manager-7765d96ddf-qbp4s\" (UID: \"305a708b-69fc-4263-a28f-015fe16402f7\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619057 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phs2d\" (UniqueName: \"kubernetes.io/projected/c7e152eb-ab3c-427a-a12b-70f46b881eee-kube-api-access-phs2d\") pod \"ovn-operator-controller-manager-b6456fdb6-68hlj\" (UID: \"c7e152eb-ab3c-427a-a12b-70f46b881eee\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619369 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqk6d\" (UniqueName: \"kubernetes.io/projected/b8304681-abf0-4e5c-aa37-41453f6e7731-kube-api-access-cqk6d\") pod \"nova-operator-controller-manager-697bc559fc-hm2q2\" (UID: \"b8304681-abf0-4e5c-aa37-41453f6e7731\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619394 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgmcr\" (UniqueName: \"kubernetes.io/projected/dbeb21cb-a705-464d-ace4-90e1ce5084a5-kube-api-access-xgmcr\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619416 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmj8n\" (UniqueName: \"kubernetes.io/projected/6925aa7e-915e-4bbf-a3e1-991ce581b49c-kube-api-access-hmj8n\") pod \"placement-operator-controller-manager-78f8948974-hj6qn\" (UID: \"6925aa7e-915e-4bbf-a3e1-991ce581b49c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619439 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrtr8\" (UniqueName: \"kubernetes.io/projected/594f9dc7-fb50-4241-bb93-90d8039339dc-kube-api-access-vrtr8\") pod \"octavia-operator-controller-manager-998648c74-9ch2g\" (UID: \"594f9dc7-fb50-4241-bb93-90d8039339dc\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619475 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619554 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgthr\" (UniqueName: \"kubernetes.io/projected/f621de25-14cd-43f5-a0a8-cf0866337715-kube-api-access-lgthr\") pod \"swift-operator-controller-manager-5f8c65bbfc-252wh\" (UID: \"f621de25-14cd-43f5-a0a8-cf0866337715\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619580 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68rmm\" (UniqueName: \"kubernetes.io/projected/3091b003-a466-4d34-8a26-d3e68a3639a9-kube-api-access-68rmm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-gv4jj\" (UID: \"3091b003-a466-4d34-8a26-d3e68a3639a9\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.619604 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlgv4\" (UniqueName: \"kubernetes.io/projected/c6d7646e-fb4a-4199-aadb-670000485477-kube-api-access-jlgv4\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hqgrk\" (UID: \"c6d7646e-fb4a-4199-aadb-670000485477\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.638838 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.645310 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlgv4\" (UniqueName: \"kubernetes.io/projected/c6d7646e-fb4a-4199-aadb-670000485477-kube-api-access-jlgv4\") pod \"mariadb-operator-controller-manager-56bbcc9d85-hqgrk\" (UID: \"c6d7646e-fb4a-4199-aadb-670000485477\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.649346 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68rmm\" (UniqueName: \"kubernetes.io/projected/3091b003-a466-4d34-8a26-d3e68a3639a9-kube-api-access-68rmm\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-gv4jj\" (UID: \"3091b003-a466-4d34-8a26-d3e68a3639a9\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.650997 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqk6d\" (UniqueName: \"kubernetes.io/projected/b8304681-abf0-4e5c-aa37-41453f6e7731-kube-api-access-cqk6d\") pod \"nova-operator-controller-manager-697bc559fc-hm2q2\" (UID: \"b8304681-abf0-4e5c-aa37-41453f6e7731\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.654854 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.661756 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.663119 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.667014 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-78l6k" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.697439 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.716108 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.717194 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.719054 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-c2g8n" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.721031 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.721123 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.721155 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgthr\" (UniqueName: \"kubernetes.io/projected/f621de25-14cd-43f5-a0a8-cf0866337715-kube-api-access-lgthr\") pod \"swift-operator-controller-manager-5f8c65bbfc-252wh\" (UID: \"f621de25-14cd-43f5-a0a8-cf0866337715\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.721237 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phs2d\" (UniqueName: \"kubernetes.io/projected/c7e152eb-ab3c-427a-a12b-70f46b881eee-kube-api-access-phs2d\") pod \"ovn-operator-controller-manager-b6456fdb6-68hlj\" (UID: \"c7e152eb-ab3c-427a-a12b-70f46b881eee\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.721260 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgmcr\" (UniqueName: \"kubernetes.io/projected/dbeb21cb-a705-464d-ace4-90e1ce5084a5-kube-api-access-xgmcr\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.721278 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmj8n\" (UniqueName: \"kubernetes.io/projected/6925aa7e-915e-4bbf-a3e1-991ce581b49c-kube-api-access-hmj8n\") pod \"placement-operator-controller-manager-78f8948974-hj6qn\" (UID: \"6925aa7e-915e-4bbf-a3e1-991ce581b49c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" Dec 03 22:12:48 crc kubenswrapper[4715]: E1203 22:12:48.721293 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:48 crc kubenswrapper[4715]: E1203 22:12:48.721356 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert podName:dbeb21cb-a705-464d-ace4-90e1ce5084a5 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:49.221339798 +0000 UTC m=+1865.964050393 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" (UID: "dbeb21cb-a705-464d-ace4-90e1ce5084a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.721298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrtr8\" (UniqueName: \"kubernetes.io/projected/594f9dc7-fb50-4241-bb93-90d8039339dc-kube-api-access-vrtr8\") pod \"octavia-operator-controller-manager-998648c74-9ch2g\" (UID: \"594f9dc7-fb50-4241-bb93-90d8039339dc\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.721524 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64gbq\" (UniqueName: \"kubernetes.io/projected/85bafa11-dcf4-4ea4-b68e-e2d4799e813d-kube-api-access-64gbq\") pod \"telemetry-operator-controller-manager-76cc84c6bb-8g9vt\" (UID: \"85bafa11-dcf4-4ea4-b68e-e2d4799e813d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.723586 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.746099 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgthr\" (UniqueName: \"kubernetes.io/projected/f621de25-14cd-43f5-a0a8-cf0866337715-kube-api-access-lgthr\") pod \"swift-operator-controller-manager-5f8c65bbfc-252wh\" (UID: \"f621de25-14cd-43f5-a0a8-cf0866337715\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.751440 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmj8n\" (UniqueName: \"kubernetes.io/projected/6925aa7e-915e-4bbf-a3e1-991ce581b49c-kube-api-access-hmj8n\") pod \"placement-operator-controller-manager-78f8948974-hj6qn\" (UID: \"6925aa7e-915e-4bbf-a3e1-991ce581b49c\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.761473 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrtr8\" (UniqueName: \"kubernetes.io/projected/594f9dc7-fb50-4241-bb93-90d8039339dc-kube-api-access-vrtr8\") pod \"octavia-operator-controller-manager-998648c74-9ch2g\" (UID: \"594f9dc7-fb50-4241-bb93-90d8039339dc\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.762332 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phs2d\" (UniqueName: \"kubernetes.io/projected/c7e152eb-ab3c-427a-a12b-70f46b881eee-kube-api-access-phs2d\") pod \"ovn-operator-controller-manager-b6456fdb6-68hlj\" (UID: \"c7e152eb-ab3c-427a-a12b-70f46b881eee\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.767838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgmcr\" (UniqueName: \"kubernetes.io/projected/dbeb21cb-a705-464d-ace4-90e1ce5084a5-kube-api-access-xgmcr\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.782407 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.807797 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.823719 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.824287 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64gbq\" (UniqueName: \"kubernetes.io/projected/85bafa11-dcf4-4ea4-b68e-e2d4799e813d-kube-api-access-64gbq\") pod \"telemetry-operator-controller-manager-76cc84c6bb-8g9vt\" (UID: \"85bafa11-dcf4-4ea4-b68e-e2d4799e813d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.824352 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzhqw\" (UniqueName: \"kubernetes.io/projected/959f935f-f997-47bc-8798-d54f7b62f2c9-kube-api-access-xzhqw\") pod \"watcher-operator-controller-manager-769dc69bc-gtpjt\" (UID: \"959f935f-f997-47bc-8798-d54f7b62f2c9\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.824439 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d57s6\" (UniqueName: \"kubernetes.io/projected/0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e-kube-api-access-d57s6\") pod \"test-operator-controller-manager-5854674fcc-ttfdp\" (UID: \"0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.824719 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.825944 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.828754 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.828955 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-gxcsv" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.830747 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.833378 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.843897 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64gbq\" (UniqueName: \"kubernetes.io/projected/85bafa11-dcf4-4ea4-b68e-e2d4799e813d-kube-api-access-64gbq\") pod \"telemetry-operator-controller-manager-76cc84c6bb-8g9vt\" (UID: \"85bafa11-dcf4-4ea4-b68e-e2d4799e813d\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.845877 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.876940 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.920315 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.921328 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.922745 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx"] Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.923292 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-bk4fx" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.924028 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.925959 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.926027 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzhqw\" (UniqueName: \"kubernetes.io/projected/959f935f-f997-47bc-8798-d54f7b62f2c9-kube-api-access-xzhqw\") pod \"watcher-operator-controller-manager-769dc69bc-gtpjt\" (UID: \"959f935f-f997-47bc-8798-d54f7b62f2c9\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.926103 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.926149 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d57s6\" (UniqueName: \"kubernetes.io/projected/0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e-kube-api-access-d57s6\") pod \"test-operator-controller-manager-5854674fcc-ttfdp\" (UID: \"0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.926177 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dggm2\" (UniqueName: \"kubernetes.io/projected/a41a8afb-d557-4caa-bf3e-57a9a5115f57-kube-api-access-dggm2\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.949114 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d57s6\" (UniqueName: \"kubernetes.io/projected/0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e-kube-api-access-d57s6\") pod \"test-operator-controller-manager-5854674fcc-ttfdp\" (UID: \"0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.952913 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.954930 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.959839 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzhqw\" (UniqueName: \"kubernetes.io/projected/959f935f-f997-47bc-8798-d54f7b62f2c9-kube-api-access-xzhqw\") pod \"watcher-operator-controller-manager-769dc69bc-gtpjt\" (UID: \"959f935f-f997-47bc-8798-d54f7b62f2c9\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" Dec 03 22:12:48 crc kubenswrapper[4715]: I1203 22:12:48.974312 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.027731 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.028181 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.028220 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfljf\" (UniqueName: \"kubernetes.io/projected/1d2cfa37-465d-4b27-8a8b-90038e1f4bf1-kube-api-access-rfljf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-hnzzx\" (UID: \"1d2cfa37-465d-4b27-8a8b-90038e1f4bf1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.028241 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.028320 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dggm2\" (UniqueName: \"kubernetes.io/projected/a41a8afb-d557-4caa-bf3e-57a9a5115f57-kube-api-access-dggm2\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.028898 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.028942 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:49.528926821 +0000 UTC m=+1866.271637406 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "metrics-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.029102 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.029154 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert podName:6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef nodeName:}" failed. No retries permitted until 2025-12-03 22:12:50.029117116 +0000 UTC m=+1866.771827711 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert") pod "infra-operator-controller-manager-57548d458d-kwnlp" (UID: "6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef") : secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.029199 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.029218 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:49.529211878 +0000 UTC m=+1866.271922473 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "webhook-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.050449 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dggm2\" (UniqueName: \"kubernetes.io/projected/a41a8afb-d557-4caa-bf3e-57a9a5115f57-kube-api-access-dggm2\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.132232 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfljf\" (UniqueName: \"kubernetes.io/projected/1d2cfa37-465d-4b27-8a8b-90038e1f4bf1-kube-api-access-rfljf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-hnzzx\" (UID: \"1d2cfa37-465d-4b27-8a8b-90038e1f4bf1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.153405 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfljf\" (UniqueName: \"kubernetes.io/projected/1d2cfa37-465d-4b27-8a8b-90038e1f4bf1-kube-api-access-rfljf\") pod \"rabbitmq-cluster-operator-manager-668c99d594-hnzzx\" (UID: \"1d2cfa37-465d-4b27-8a8b-90038e1f4bf1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.233889 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.234028 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.234081 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert podName:dbeb21cb-a705-464d-ace4-90e1ce5084a5 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:50.234064826 +0000 UTC m=+1866.976775421 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" (UID: "dbeb21cb-a705-464d-ace4-90e1ce5084a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.316248 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.417854 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86"] Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.430444 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8cbfb12_17c2_46bf_baea_ee8ac16db4ee.slice/crio-993bc73c50838fc3f8872c48b8de8beac150e4601c2b1e61aa3dbde02d6a356d WatchSource:0}: Error finding container 993bc73c50838fc3f8872c48b8de8beac150e4601c2b1e61aa3dbde02d6a356d: Status 404 returned error can't find the container with id 993bc73c50838fc3f8872c48b8de8beac150e4601c2b1e61aa3dbde02d6a356d Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.430880 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr"] Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.447509 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb"] Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.452348 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9"] Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.462161 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx"] Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.465561 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn"] Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.537213 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.537600 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.537769 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.537824 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:50.537801368 +0000 UTC m=+1867.280511963 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "metrics-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.538190 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.538221 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:50.538210429 +0000 UTC m=+1867.280921034 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "webhook-server-cert" not found Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.595036 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m"] Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.769097 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s"] Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.781069 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod305a708b_69fc_4263_a28f_015fe16402f7.slice/crio-391623eb56f0167c0dccfda07a6de4eb901ffb4f49bf1f51296cedf021b12773 WatchSource:0}: Error finding container 391623eb56f0167c0dccfda07a6de4eb901ffb4f49bf1f51296cedf021b12773: Status 404 returned error can't find the container with id 391623eb56f0167c0dccfda07a6de4eb901ffb4f49bf1f51296cedf021b12773 Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.795468 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk"] Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.803730 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6d7646e_fb4a_4199_aadb_670000485477.slice/crio-e8abea96176a51cb4115e6994d5fdd361b0f1317bd9b275cd9601d65c4c4c3c9 WatchSource:0}: Error finding container e8abea96176a51cb4115e6994d5fdd361b0f1317bd9b275cd9601d65c4c4c3c9: Status 404 returned error can't find the container with id e8abea96176a51cb4115e6994d5fdd361b0f1317bd9b275cd9601d65c4c4c3c9 Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.806298 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh"] Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.814438 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2"] Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.814573 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3091b003_a466_4d34_8a26_d3e68a3639a9.slice/crio-2b5fa48b927d3aa31e06826ef37e1a6ef7d71940ea1507d4678b2cb95a0064a9 WatchSource:0}: Error finding container 2b5fa48b927d3aa31e06826ef37e1a6ef7d71940ea1507d4678b2cb95a0064a9: Status 404 returned error can't find the container with id 2b5fa48b927d3aa31e06826ef37e1a6ef7d71940ea1507d4678b2cb95a0064a9 Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.817974 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95b90127_8325_43c8_977e_8b972b62db2a.slice/crio-954e9bc0e3ac929b6c406c5e4f1bd1447ded03aa2b1947bf5f045ad1ed966379 WatchSource:0}: Error finding container 954e9bc0e3ac929b6c406c5e4f1bd1447ded03aa2b1947bf5f045ad1ed966379: Status 404 returned error can't find the container with id 954e9bc0e3ac929b6c406c5e4f1bd1447ded03aa2b1947bf5f045ad1ed966379 Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.825585 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb8304681_abf0_4e5c_aa37_41453f6e7731.slice/crio-8f6541d2a06e8d36a6eb35a88d45da121e2c5a12abfb9bd6bf50756d38891bd2 WatchSource:0}: Error finding container 8f6541d2a06e8d36a6eb35a88d45da121e2c5a12abfb9bd6bf50756d38891bd2: Status 404 returned error can't find the container with id 8f6541d2a06e8d36a6eb35a88d45da121e2c5a12abfb9bd6bf50756d38891bd2 Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.830643 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj"] Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.830953 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cqk6d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-hm2q2_openstack-operators(b8304681-abf0-4e5c-aa37-41453f6e7731): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.837835 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d4b2d7c_f73c_4ab9_a5e1_775a99673f7e.slice/crio-116269ae009d85dd6fa8ac961b3a45eb04b72872f1b1cf953c9b873a674b392e WatchSource:0}: Error finding container 116269ae009d85dd6fa8ac961b3a45eb04b72872f1b1cf953c9b873a674b392e: Status 404 returned error can't find the container with id 116269ae009d85dd6fa8ac961b3a45eb04b72872f1b1cf953c9b873a674b392e Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.839882 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf621de25_14cd_43f5_a0a8_cf0866337715.slice/crio-c2ab35807e368bf9d5c7113b5e52c11e9e66624275983af100e3703535e395d9 WatchSource:0}: Error finding container c2ab35807e368bf9d5c7113b5e52c11e9e66624275983af100e3703535e395d9: Status 404 returned error can't find the container with id c2ab35807e368bf9d5c7113b5e52c11e9e66624275983af100e3703535e395d9 Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.853602 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp"] Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.860544 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt"] Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.861832 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lgthr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-252wh_openstack-operators(f621de25-14cd-43f5-a0a8-cf0866337715): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.863291 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d57s6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-ttfdp_openstack-operators(0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.864223 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lgthr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-252wh_openstack-operators(f621de25-14cd-43f5-a0a8-cf0866337715): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.864272 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh"] Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.866795 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" podUID="f621de25-14cd-43f5-a0a8-cf0866337715" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.866923 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d57s6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-ttfdp_openstack-operators(0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.868929 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" podUID="0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e" Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.869461 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod594f9dc7_fb50_4241_bb93_90d8039339dc.slice/crio-59f2fc4b8236b179bcc42dbb16435872eea3539acbdb1e2b164bff5f1e1b0896 WatchSource:0}: Error finding container 59f2fc4b8236b179bcc42dbb16435872eea3539acbdb1e2b164bff5f1e1b0896: Status 404 returned error can't find the container with id 59f2fc4b8236b179bcc42dbb16435872eea3539acbdb1e2b164bff5f1e1b0896 Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.871356 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj"] Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.872513 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod959f935f_f997_47bc_8798_d54f7b62f2c9.slice/crio-82278d4e1fb3ed7c04917cf05562501bfd64331ff3900bda661895018067f7a1 WatchSource:0}: Error finding container 82278d4e1fb3ed7c04917cf05562501bfd64331ff3900bda661895018067f7a1: Status 404 returned error can't find the container with id 82278d4e1fb3ed7c04917cf05562501bfd64331ff3900bda661895018067f7a1 Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.873988 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod85bafa11_dcf4_4ea4_b68e_e2d4799e813d.slice/crio-99f3523dbcf0e9efa9c3f3f07444d3573faecbdbb82507a62a44f5cf65d26e06 WatchSource:0}: Error finding container 99f3523dbcf0e9efa9c3f3f07444d3573faecbdbb82507a62a44f5cf65d26e06: Status 404 returned error can't find the container with id 99f3523dbcf0e9efa9c3f3f07444d3573faecbdbb82507a62a44f5cf65d26e06 Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.873960 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vrtr8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-9ch2g_openstack-operators(594f9dc7-fb50-4241-bb93-90d8039339dc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.876689 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt"] Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.877257 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vrtr8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-9ch2g_openstack-operators(594f9dc7-fb50-4241-bb93-90d8039339dc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.878438 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" podUID="594f9dc7-fb50-4241-bb93-90d8039339dc" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.880484 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-64gbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-8g9vt_openstack-operators(85bafa11-dcf4-4ea4-b68e-e2d4799e813d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.880845 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xzhqw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-gtpjt_openstack-operators(959f935f-f997-47bc-8798-d54f7b62f2c9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.882051 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-phs2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-68hlj_openstack-operators(c7e152eb-ab3c-427a-a12b-70f46b881eee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.882153 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-64gbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-8g9vt_openstack-operators(85bafa11-dcf4-4ea4-b68e-e2d4799e813d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.883135 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xzhqw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-gtpjt_openstack-operators(959f935f-f997-47bc-8798-d54f7b62f2c9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.883466 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" podUID="85bafa11-dcf4-4ea4-b68e-e2d4799e813d" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.884037 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-phs2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-68hlj_openstack-operators(c7e152eb-ab3c-427a-a12b-70f46b881eee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.884099 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn"] Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.885842 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" podUID="c7e152eb-ab3c-427a-a12b-70f46b881eee" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.886125 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" podUID="959f935f-f997-47bc-8798-d54f7b62f2c9" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.889698 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g"] Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.890210 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmj8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-hj6qn_openstack-operators(6925aa7e-915e-4bbf-a3e1-991ce581b49c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: I1203 22:12:49.893647 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx"] Dec 03 22:12:49 crc kubenswrapper[4715]: W1203 22:12:49.896074 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1d2cfa37_465d_4b27_8a8b_90038e1f4bf1.slice/crio-d70d267721b87eb6693ff98c2a3c7c83d9f714dcb3d6fe837e3d079eab74e058 WatchSource:0}: Error finding container d70d267721b87eb6693ff98c2a3c7c83d9f714dcb3d6fe837e3d079eab74e058: Status 404 returned error can't find the container with id d70d267721b87eb6693ff98c2a3c7c83d9f714dcb3d6fe837e3d079eab74e058 Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.898611 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hmj8n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-hj6qn_openstack-operators(6925aa7e-915e-4bbf-a3e1-991ce581b49c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.899835 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" podUID="6925aa7e-915e-4bbf-a3e1-991ce581b49c" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.900959 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rfljf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-hnzzx_openstack-operators(1d2cfa37-465d-4b27-8a8b-90038e1f4bf1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 22:12:49 crc kubenswrapper[4715]: E1203 22:12:49.902122 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" podUID="1d2cfa37-465d-4b27-8a8b-90038e1f4bf1" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.047843 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.048057 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.048199 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert podName:6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef nodeName:}" failed. No retries permitted until 2025-12-03 22:12:52.048154864 +0000 UTC m=+1868.790865459 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert") pod "infra-operator-controller-manager-57548d458d-kwnlp" (UID: "6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef") : secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.134177 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" event={"ID":"9c8c4875-3433-4885-9afa-acbfea9d546b","Type":"ContainerStarted","Data":"e9cf7cb99e2b97d8fb49aa2888eda9dc68b31c3fe9c6f418c0bd0bd597723d5e"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.135352 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" event={"ID":"594f9dc7-fb50-4241-bb93-90d8039339dc","Type":"ContainerStarted","Data":"59f2fc4b8236b179bcc42dbb16435872eea3539acbdb1e2b164bff5f1e1b0896"} Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.139674 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" podUID="594f9dc7-fb50-4241-bb93-90d8039339dc" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.140124 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" event={"ID":"305a708b-69fc-4263-a28f-015fe16402f7","Type":"ContainerStarted","Data":"391623eb56f0167c0dccfda07a6de4eb901ffb4f49bf1f51296cedf021b12773"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.141940 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" event={"ID":"c6d7646e-fb4a-4199-aadb-670000485477","Type":"ContainerStarted","Data":"e8abea96176a51cb4115e6994d5fdd361b0f1317bd9b275cd9601d65c4c4c3c9"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.144772 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" event={"ID":"95b90127-8325-43c8-977e-8b972b62db2a","Type":"ContainerStarted","Data":"954e9bc0e3ac929b6c406c5e4f1bd1447ded03aa2b1947bf5f045ad1ed966379"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.148714 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" event={"ID":"3091b003-a466-4d34-8a26-d3e68a3639a9","Type":"ContainerStarted","Data":"2b5fa48b927d3aa31e06826ef37e1a6ef7d71940ea1507d4678b2cb95a0064a9"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.152022 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" event={"ID":"1d2cfa37-465d-4b27-8a8b-90038e1f4bf1","Type":"ContainerStarted","Data":"d70d267721b87eb6693ff98c2a3c7c83d9f714dcb3d6fe837e3d079eab74e058"} Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.153530 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" podUID="1d2cfa37-465d-4b27-8a8b-90038e1f4bf1" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.155151 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" event={"ID":"46e49f12-51e7-45ca-a0cb-c7f8a40adf12","Type":"ContainerStarted","Data":"3bc1426c2504d307f8e4208a41be337a1e435f250f3d24e535cc368040dd695f"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.161086 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" event={"ID":"b8cbfb12-17c2-46bf-baea-ee8ac16db4ee","Type":"ContainerStarted","Data":"993bc73c50838fc3f8872c48b8de8beac150e4601c2b1e61aa3dbde02d6a356d"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.164461 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" event={"ID":"8219810d-511e-4cc8-a018-49f425b339c0","Type":"ContainerStarted","Data":"52d43274ef89d324844d9a4547fe64c340e93f2b7cb1c66741294c9035bf68f5"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.168447 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" event={"ID":"9a900ae2-08bc-403b-a613-52e7870eeebf","Type":"ContainerStarted","Data":"8fea05f3dc851188f62bd4ce43cd4dc082d2be89233da9c7980d45044c584c47"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.174327 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"2a7b507f93c05d0aab4e197e7602419073e36d87d0fa5a60853ef24a637d4567"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.200141 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" event={"ID":"b8304681-abf0-4e5c-aa37-41453f6e7731","Type":"ContainerStarted","Data":"8f6541d2a06e8d36a6eb35a88d45da121e2c5a12abfb9bd6bf50756d38891bd2"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.208483 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" event={"ID":"a91ec3b7-705a-4500-be8e-e8c789a2c973","Type":"ContainerStarted","Data":"f37a96113382f808c772e6ebcd079f44fa5f95f51d21900f8b2ff9c55b0a04e4"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.211650 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" event={"ID":"1cba7fbc-0a66-45b4-b20d-33c81d407b96","Type":"ContainerStarted","Data":"7f1b9d39664df86bb96b745b11002198274a7ec2079f8ad385c635f5a3f5c442"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.214667 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" event={"ID":"c7e152eb-ab3c-427a-a12b-70f46b881eee","Type":"ContainerStarted","Data":"3eaeb9404067458fec57a665ba00bdf2a6db03e9ea3b9a5726e9d89b4fd58bce"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.217028 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" event={"ID":"6925aa7e-915e-4bbf-a3e1-991ce581b49c","Type":"ContainerStarted","Data":"5c04420ec2941cf55260dfc6784d2bcd88391cc2b92548bebe3fc704f9eb3917"} Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.218229 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" podUID="c7e152eb-ab3c-427a-a12b-70f46b881eee" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.218868 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" event={"ID":"f621de25-14cd-43f5-a0a8-cf0866337715","Type":"ContainerStarted","Data":"c2ab35807e368bf9d5c7113b5e52c11e9e66624275983af100e3703535e395d9"} Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.220404 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" podUID="6925aa7e-915e-4bbf-a3e1-991ce581b49c" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.230645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" event={"ID":"85bafa11-dcf4-4ea4-b68e-e2d4799e813d","Type":"ContainerStarted","Data":"99f3523dbcf0e9efa9c3f3f07444d3573faecbdbb82507a62a44f5cf65d26e06"} Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.235119 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" podUID="f621de25-14cd-43f5-a0a8-cf0866337715" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.238121 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" event={"ID":"959f935f-f997-47bc-8798-d54f7b62f2c9","Type":"ContainerStarted","Data":"82278d4e1fb3ed7c04917cf05562501bfd64331ff3900bda661895018067f7a1"} Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.240187 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" podUID="85bafa11-dcf4-4ea4-b68e-e2d4799e813d" Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.240792 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" podUID="959f935f-f997-47bc-8798-d54f7b62f2c9" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.250230 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" event={"ID":"0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e","Type":"ContainerStarted","Data":"116269ae009d85dd6fa8ac961b3a45eb04b72872f1b1cf953c9b873a674b392e"} Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.250327 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.250867 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.250929 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert podName:dbeb21cb-a705-464d-ace4-90e1ce5084a5 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:52.250914546 +0000 UTC m=+1868.993625141 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" (UID: "dbeb21cb-a705-464d-ace4-90e1ce5084a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.260462 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" podUID="0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.554646 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:50 crc kubenswrapper[4715]: I1203 22:12:50.555061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.554826 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.555272 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:52.555253784 +0000 UTC m=+1869.297964389 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "webhook-server-cert" not found Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.555210 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 22:12:50 crc kubenswrapper[4715]: E1203 22:12:50.555411 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:52.555391857 +0000 UTC m=+1869.298102452 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "metrics-server-cert" not found Dec 03 22:12:51 crc kubenswrapper[4715]: E1203 22:12:51.264649 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" podUID="6925aa7e-915e-4bbf-a3e1-991ce581b49c" Dec 03 22:12:51 crc kubenswrapper[4715]: E1203 22:12:51.264884 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" podUID="c7e152eb-ab3c-427a-a12b-70f46b881eee" Dec 03 22:12:51 crc kubenswrapper[4715]: E1203 22:12:51.265073 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" podUID="1d2cfa37-465d-4b27-8a8b-90038e1f4bf1" Dec 03 22:12:51 crc kubenswrapper[4715]: E1203 22:12:51.265132 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" podUID="959f935f-f997-47bc-8798-d54f7b62f2c9" Dec 03 22:12:51 crc kubenswrapper[4715]: E1203 22:12:51.266272 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:d9a3694865a7d54ee96397add18c3898886e98d079aa20876a0f4de1fa7a7168\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" podUID="594f9dc7-fb50-4241-bb93-90d8039339dc" Dec 03 22:12:51 crc kubenswrapper[4715]: E1203 22:12:51.266269 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" podUID="0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e" Dec 03 22:12:51 crc kubenswrapper[4715]: E1203 22:12:51.266313 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" podUID="f621de25-14cd-43f5-a0a8-cf0866337715" Dec 03 22:12:51 crc kubenswrapper[4715]: E1203 22:12:51.266342 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" podUID="85bafa11-dcf4-4ea4-b68e-e2d4799e813d" Dec 03 22:12:52 crc kubenswrapper[4715]: I1203 22:12:52.077357 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:52 crc kubenswrapper[4715]: E1203 22:12:52.077566 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:52 crc kubenswrapper[4715]: E1203 22:12:52.077749 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert podName:6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef nodeName:}" failed. No retries permitted until 2025-12-03 22:12:56.077734076 +0000 UTC m=+1872.820444671 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert") pod "infra-operator-controller-manager-57548d458d-kwnlp" (UID: "6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef") : secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:52 crc kubenswrapper[4715]: I1203 22:12:52.283380 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:52 crc kubenswrapper[4715]: E1203 22:12:52.283596 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:52 crc kubenswrapper[4715]: E1203 22:12:52.283678 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert podName:dbeb21cb-a705-464d-ace4-90e1ce5084a5 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:56.283656822 +0000 UTC m=+1873.026367427 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" (UID: "dbeb21cb-a705-464d-ace4-90e1ce5084a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:52 crc kubenswrapper[4715]: I1203 22:12:52.586496 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:52 crc kubenswrapper[4715]: I1203 22:12:52.586940 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:52 crc kubenswrapper[4715]: E1203 22:12:52.586770 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 22:12:52 crc kubenswrapper[4715]: E1203 22:12:52.587085 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 22:12:52 crc kubenswrapper[4715]: E1203 22:12:52.587104 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:56.587071336 +0000 UTC m=+1873.329782011 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "webhook-server-cert" not found Dec 03 22:12:52 crc kubenswrapper[4715]: E1203 22:12:52.587140 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:12:56.587124227 +0000 UTC m=+1873.329834822 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "metrics-server-cert" not found Dec 03 22:12:56 crc kubenswrapper[4715]: I1203 22:12:56.141324 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:12:56 crc kubenswrapper[4715]: E1203 22:12:56.141593 4715 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:56 crc kubenswrapper[4715]: E1203 22:12:56.141850 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert podName:6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef nodeName:}" failed. No retries permitted until 2025-12-03 22:13:04.1418286 +0000 UTC m=+1880.884539195 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert") pod "infra-operator-controller-manager-57548d458d-kwnlp" (UID: "6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef") : secret "infra-operator-webhook-server-cert" not found Dec 03 22:12:56 crc kubenswrapper[4715]: I1203 22:12:56.344495 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:12:56 crc kubenswrapper[4715]: E1203 22:12:56.344623 4715 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:56 crc kubenswrapper[4715]: E1203 22:12:56.344686 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert podName:dbeb21cb-a705-464d-ace4-90e1ce5084a5 nodeName:}" failed. No retries permitted until 2025-12-03 22:13:04.344667914 +0000 UTC m=+1881.087378509 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" (UID: "dbeb21cb-a705-464d-ace4-90e1ce5084a5") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 22:12:56 crc kubenswrapper[4715]: I1203 22:12:56.648136 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:56 crc kubenswrapper[4715]: I1203 22:12:56.648255 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:12:56 crc kubenswrapper[4715]: E1203 22:12:56.648307 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 22:12:56 crc kubenswrapper[4715]: E1203 22:12:56.648390 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:13:04.648375276 +0000 UTC m=+1881.391085871 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "metrics-server-cert" not found Dec 03 22:12:56 crc kubenswrapper[4715]: E1203 22:12:56.648388 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 22:12:56 crc kubenswrapper[4715]: E1203 22:12:56.648439 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:13:04.648425827 +0000 UTC m=+1881.391136422 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "webhook-server-cert" not found Dec 03 22:13:02 crc kubenswrapper[4715]: E1203 22:13:02.018051 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 22:13:02 crc kubenswrapper[4715]: E1203 22:13:02.018721 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2mxtk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-qbp4s_openstack-operators(305a708b-69fc-4263-a28f-015fe16402f7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:13:02 crc kubenswrapper[4715]: E1203 22:13:02.697346 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 22:13:02 crc kubenswrapper[4715]: E1203 22:13:02.697488 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cqk6d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-hm2q2_openstack-operators(b8304681-abf0-4e5c-aa37-41453f6e7731): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:13:02 crc kubenswrapper[4715]: E1203 22:13:02.698694 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" podUID="b8304681-abf0-4e5c-aa37-41453f6e7731" Dec 03 22:13:03 crc kubenswrapper[4715]: E1203 22:13:03.187354 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" podUID="305a708b-69fc-4263-a28f-015fe16402f7" Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.370857 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" event={"ID":"9c8c4875-3433-4885-9afa-acbfea9d546b","Type":"ContainerStarted","Data":"400f949916eb2d2a35a4e94e49ada0c6364e5e6ab80ba2385cf811a5c53e4fb5"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.389690 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" event={"ID":"1cba7fbc-0a66-45b4-b20d-33c81d407b96","Type":"ContainerStarted","Data":"e0c831e297771af88b83af78b2bcacb32ad763ef828520e29648b9545198bbe0"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.392944 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" event={"ID":"95b90127-8325-43c8-977e-8b972b62db2a","Type":"ContainerStarted","Data":"4b9edea0fc23484b55ae901e76aab35637850a6c076f20086d4134561eaf757c"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.394105 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" event={"ID":"3091b003-a466-4d34-8a26-d3e68a3639a9","Type":"ContainerStarted","Data":"a07933e2f4f562260f45be05ea0cb050424f7d82a181a9b5148f1326a58f7ad8"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.395478 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" event={"ID":"9a900ae2-08bc-403b-a613-52e7870eeebf","Type":"ContainerStarted","Data":"75fa835569567988a593cb19f75cadc2f78f25e228b718db66aa25d8aa702e37"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.397699 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" event={"ID":"46e49f12-51e7-45ca-a0cb-c7f8a40adf12","Type":"ContainerStarted","Data":"fb5f87c63854d3f48e98aa11bb67117ae0a30341eaeb26f708dbd6204c5c1df9"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.404094 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" event={"ID":"8219810d-511e-4cc8-a018-49f425b339c0","Type":"ContainerStarted","Data":"bf55f070b15dcdc172d4a0b6efa81cc78ad7b9bc8852dfce1bee1984ca518ac1"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.407677 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" event={"ID":"305a708b-69fc-4263-a28f-015fe16402f7","Type":"ContainerStarted","Data":"b24ed519bcce25cb1778901379183ccb531095941d2da2f729c36b059fe1e34e"} Dec 03 22:13:03 crc kubenswrapper[4715]: E1203 22:13:03.409698 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" podUID="305a708b-69fc-4263-a28f-015fe16402f7" Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.414747 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" event={"ID":"c6d7646e-fb4a-4199-aadb-670000485477","Type":"ContainerStarted","Data":"32999d0017c1c8892ca0007755a6e3e134eebcbd99b903fb571f700ab077e5a1"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.416383 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" event={"ID":"a91ec3b7-705a-4500-be8e-e8c789a2c973","Type":"ContainerStarted","Data":"ee05f511f806c648d4feb794dd08e943bba68e9178694f1463cfa2b8ec29c865"} Dec 03 22:13:03 crc kubenswrapper[4715]: I1203 22:13:03.421434 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" event={"ID":"b8cbfb12-17c2-46bf-baea-ee8ac16db4ee","Type":"ContainerStarted","Data":"a1d81fff6006a4bf51417f4c8c93a60ef311d95f14d343ffe727157052ca3d88"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.200787 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.223387 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef-cert\") pod \"infra-operator-controller-manager-57548d458d-kwnlp\" (UID: \"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.403685 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.407220 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dbeb21cb-a705-464d-ace4-90e1ce5084a5-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp\" (UID: \"dbeb21cb-a705-464d-ace4-90e1ce5084a5\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.443130 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" event={"ID":"c6d7646e-fb4a-4199-aadb-670000485477","Type":"ContainerStarted","Data":"d8e2743dfbcf108d89e45ae3e0bcb5560587ee919eb265af52d06b160aa684fe"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.443406 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.452088 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" event={"ID":"a91ec3b7-705a-4500-be8e-e8c789a2c973","Type":"ContainerStarted","Data":"ea2fb56145af650decb53eb7df9541d2531eb7e43c217f36232fb45b241cf9f3"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.452144 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.458667 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" event={"ID":"b8cbfb12-17c2-46bf-baea-ee8ac16db4ee","Type":"ContainerStarted","Data":"48db794e344a68dea7e2929bd8852909c9175ffd274e52f3319821c4d57f17e8"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.458969 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" podStartSLOduration=3.5842457850000002 podStartE2EDuration="16.458959004s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.811945621 +0000 UTC m=+1866.554656216" lastFinishedPulling="2025-12-03 22:13:02.68665882 +0000 UTC m=+1879.429369435" observedRunningTime="2025-12-03 22:13:04.458848991 +0000 UTC m=+1881.201559586" watchObservedRunningTime="2025-12-03 22:13:04.458959004 +0000 UTC m=+1881.201669589" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.459249 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.467013 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.470888 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.474771 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" event={"ID":"9c8c4875-3433-4885-9afa-acbfea9d546b","Type":"ContainerStarted","Data":"036d72bbb819ae974d6bc40451513fa88634a4194a9520ead26bac21bc3ca828"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.475208 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.476657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" event={"ID":"1cba7fbc-0a66-45b4-b20d-33c81d407b96","Type":"ContainerStarted","Data":"9e9322f1fc6602394a0f40ea10ac4a8399a515b0c9f873b0deb91c64b64ade24"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.477004 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.486157 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" podStartSLOduration=3.847002481 podStartE2EDuration="16.4861449s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.454870053 +0000 UTC m=+1866.197580658" lastFinishedPulling="2025-12-03 22:13:02.094012442 +0000 UTC m=+1878.836723077" observedRunningTime="2025-12-03 22:13:04.484758063 +0000 UTC m=+1881.227468658" watchObservedRunningTime="2025-12-03 22:13:04.4861449 +0000 UTC m=+1881.228855495" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.501393 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" event={"ID":"8219810d-511e-4cc8-a018-49f425b339c0","Type":"ContainerStarted","Data":"cc058aaad0ad2728943bc81f9080d659680e26590c3cef92c51070749fb80aeb"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.501543 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.507658 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" podStartSLOduration=3.901035866 podStartE2EDuration="16.507643495s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.432572366 +0000 UTC m=+1866.175282971" lastFinishedPulling="2025-12-03 22:13:02.039180005 +0000 UTC m=+1878.781890600" observedRunningTime="2025-12-03 22:13:04.505851327 +0000 UTC m=+1881.248561922" watchObservedRunningTime="2025-12-03 22:13:04.507643495 +0000 UTC m=+1881.250354090" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.515605 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" event={"ID":"95b90127-8325-43c8-977e-8b972b62db2a","Type":"ContainerStarted","Data":"f69f68e3114c807da7584b6849e0d9f201461584d3fa1885b802c6229db7d55a"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.516493 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.520343 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" event={"ID":"3091b003-a466-4d34-8a26-d3e68a3639a9","Type":"ContainerStarted","Data":"6b5cfbcf353e0726108c549c439d53780b1ebafd132ebd060d8805fc0e5f7351"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.520768 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.522227 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" event={"ID":"9a900ae2-08bc-403b-a613-52e7870eeebf","Type":"ContainerStarted","Data":"0444c9e0ad80ecd5ea2966e3e83a5f5bf01506c0e5f68a65026f85d1ab73beb0"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.522632 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.542881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" event={"ID":"46e49f12-51e7-45ca-a0cb-c7f8a40adf12","Type":"ContainerStarted","Data":"f5d1243e7ae766442fe3e4151a907efe7b5600916f45bbba9f4788e8f21eea3b"} Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.542918 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" Dec 03 22:13:04 crc kubenswrapper[4715]: E1203 22:13:04.547196 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" podUID="305a708b-69fc-4263-a28f-015fe16402f7" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.554375 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" podStartSLOduration=3.930362224 podStartE2EDuration="16.554362625s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.470218317 +0000 UTC m=+1866.212928912" lastFinishedPulling="2025-12-03 22:13:02.094218678 +0000 UTC m=+1878.836929313" observedRunningTime="2025-12-03 22:13:04.52465403 +0000 UTC m=+1881.267364625" watchObservedRunningTime="2025-12-03 22:13:04.554362625 +0000 UTC m=+1881.297073220" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.554787 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" podStartSLOduration=3.9275356009999998 podStartE2EDuration="16.554783316s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.466723585 +0000 UTC m=+1866.209434180" lastFinishedPulling="2025-12-03 22:13:02.0939713 +0000 UTC m=+1878.836681895" observedRunningTime="2025-12-03 22:13:04.549564756 +0000 UTC m=+1881.292275351" watchObservedRunningTime="2025-12-03 22:13:04.554783316 +0000 UTC m=+1881.297493911" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.565053 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" podStartSLOduration=3.969592084 podStartE2EDuration="16.56504012s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.442688622 +0000 UTC m=+1866.185399227" lastFinishedPulling="2025-12-03 22:13:02.038136668 +0000 UTC m=+1878.780847263" observedRunningTime="2025-12-03 22:13:04.564717641 +0000 UTC m=+1881.307428236" watchObservedRunningTime="2025-12-03 22:13:04.56504012 +0000 UTC m=+1881.307750715" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.585109 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" podStartSLOduration=4.37340992 podStartE2EDuration="16.585095016s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.827896161 +0000 UTC m=+1866.570606756" lastFinishedPulling="2025-12-03 22:13:02.039581257 +0000 UTC m=+1878.782291852" observedRunningTime="2025-12-03 22:13:04.583295388 +0000 UTC m=+1881.326005983" watchObservedRunningTime="2025-12-03 22:13:04.585095016 +0000 UTC m=+1881.327805611" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.605618 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" podStartSLOduration=4.117842343 podStartE2EDuration="16.605603444s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.607851354 +0000 UTC m=+1866.350561949" lastFinishedPulling="2025-12-03 22:13:02.095612415 +0000 UTC m=+1878.838323050" observedRunningTime="2025-12-03 22:13:04.601902986 +0000 UTC m=+1881.344613581" watchObservedRunningTime="2025-12-03 22:13:04.605603444 +0000 UTC m=+1881.348314039" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.708770 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:04 crc kubenswrapper[4715]: E1203 22:13:04.708905 4715 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 22:13:04 crc kubenswrapper[4715]: E1203 22:13:04.709231 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:13:20.709213625 +0000 UTC m=+1897.451924220 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "webhook-server-cert" not found Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.710622 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.710795 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" podStartSLOduration=4.173170419 podStartE2EDuration="16.710780017s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.440598747 +0000 UTC m=+1866.183309342" lastFinishedPulling="2025-12-03 22:13:01.978208345 +0000 UTC m=+1878.720918940" observedRunningTime="2025-12-03 22:13:04.692057756 +0000 UTC m=+1881.434768351" watchObservedRunningTime="2025-12-03 22:13:04.710780017 +0000 UTC m=+1881.453490612" Dec 03 22:13:04 crc kubenswrapper[4715]: E1203 22:13:04.711622 4715 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 22:13:04 crc kubenswrapper[4715]: E1203 22:13:04.711699 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs podName:a41a8afb-d557-4caa-bf3e-57a9a5115f57 nodeName:}" failed. No retries permitted until 2025-12-03 22:13:20.711681421 +0000 UTC m=+1897.454392016 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs") pod "openstack-operator-controller-manager-bd9c944fb-mkhhs" (UID: "a41a8afb-d557-4caa-bf3e-57a9a5115f57") : secret "metrics-server-cert" not found Dec 03 22:13:04 crc kubenswrapper[4715]: I1203 22:13:04.714406 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" podStartSLOduration=4.50159574 podStartE2EDuration="16.714392674s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.825978261 +0000 UTC m=+1866.568688856" lastFinishedPulling="2025-12-03 22:13:02.038775195 +0000 UTC m=+1878.781485790" observedRunningTime="2025-12-03 22:13:04.672901754 +0000 UTC m=+1881.415612349" watchObservedRunningTime="2025-12-03 22:13:04.714392674 +0000 UTC m=+1881.457103269" Dec 03 22:13:05 crc kubenswrapper[4715]: I1203 22:13:05.064292 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp"] Dec 03 22:13:05 crc kubenswrapper[4715]: W1203 22:13:05.091410 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbeb21cb_a705_464d_ace4_90e1ce5084a5.slice/crio-7f1b59428f5ce0a2d6ae237a38211e04a72799b4a956e9449fcefca89103ee05 WatchSource:0}: Error finding container 7f1b59428f5ce0a2d6ae237a38211e04a72799b4a956e9449fcefca89103ee05: Status 404 returned error can't find the container with id 7f1b59428f5ce0a2d6ae237a38211e04a72799b4a956e9449fcefca89103ee05 Dec 03 22:13:05 crc kubenswrapper[4715]: W1203 22:13:05.187798 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6aea4a92_7dfd_4de9_b9a2_f504ed1ac0ef.slice/crio-0fe99ab3896208c60fe0d3b04281b804423652b0a815ea5b8dcfe58260d8e44b WatchSource:0}: Error finding container 0fe99ab3896208c60fe0d3b04281b804423652b0a815ea5b8dcfe58260d8e44b: Status 404 returned error can't find the container with id 0fe99ab3896208c60fe0d3b04281b804423652b0a815ea5b8dcfe58260d8e44b Dec 03 22:13:05 crc kubenswrapper[4715]: I1203 22:13:05.196324 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp"] Dec 03 22:13:05 crc kubenswrapper[4715]: I1203 22:13:05.551420 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" event={"ID":"dbeb21cb-a705-464d-ace4-90e1ce5084a5","Type":"ContainerStarted","Data":"7f1b59428f5ce0a2d6ae237a38211e04a72799b4a956e9449fcefca89103ee05"} Dec 03 22:13:05 crc kubenswrapper[4715]: I1203 22:13:05.554582 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" event={"ID":"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef","Type":"ContainerStarted","Data":"0fe99ab3896208c60fe0d3b04281b804423652b0a815ea5b8dcfe58260d8e44b"} Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.411535 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5hzjb" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.426461 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-mnd86" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.445454 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-sctl9" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.470604 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-d5phr" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.485176 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-gc2vn" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.547203 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-tgkbx" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.591805 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-9qbjh" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.659259 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-9gf6m" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.724235 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-hqgrk" Dec 03 22:13:08 crc kubenswrapper[4715]: I1203 22:13:08.785343 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-gv4jj" Dec 03 22:13:20 crc kubenswrapper[4715]: I1203 22:13:20.789300 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:20 crc kubenswrapper[4715]: I1203 22:13:20.790010 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:20 crc kubenswrapper[4715]: I1203 22:13:20.796789 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-webhook-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:20 crc kubenswrapper[4715]: I1203 22:13:20.797355 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a41a8afb-d557-4caa-bf3e-57a9a5115f57-metrics-certs\") pod \"openstack-operator-controller-manager-bd9c944fb-mkhhs\" (UID: \"a41a8afb-d557-4caa-bf3e-57a9a5115f57\") " pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:21 crc kubenswrapper[4715]: I1203 22:13:21.092225 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:29 crc kubenswrapper[4715]: E1203 22:13:29.541809 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7" Dec 03 22:13:29 crc kubenswrapper[4715]: E1203 22:13:29.543081 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9dfff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-kwnlp_openstack-operators(6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:13:30 crc kubenswrapper[4715]: E1203 22:13:30.385604 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621" Dec 03 22:13:30 crc kubenswrapper[4715]: E1203 22:13:30.385921 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xzhqw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-gtpjt_openstack-operators(959f935f-f997-47bc-8798-d54f7b62f2c9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:13:31 crc kubenswrapper[4715]: E1203 22:13:31.265277 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 22:13:31 crc kubenswrapper[4715]: E1203 22:13:31.265428 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rfljf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-hnzzx_openstack-operators(1d2cfa37-465d-4b27-8a8b-90038e1f4bf1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:13:31 crc kubenswrapper[4715]: E1203 22:13:31.266774 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" podUID="1d2cfa37-465d-4b27-8a8b-90038e1f4bf1" Dec 03 22:13:31 crc kubenswrapper[4715]: E1203 22:13:31.726897 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" podUID="959f935f-f997-47bc-8798-d54f7b62f2c9" Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.743290 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs"] Dec 03 22:13:31 crc kubenswrapper[4715]: E1203 22:13:31.822010 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" podUID="6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef" Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.822554 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" event={"ID":"6925aa7e-915e-4bbf-a3e1-991ce581b49c","Type":"ContainerStarted","Data":"e190efaa4b641aa98bd0ca486e215b4a5e107f6806503759e4187226f86bf5b1"} Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.841740 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" event={"ID":"305a708b-69fc-4263-a28f-015fe16402f7","Type":"ContainerStarted","Data":"46f3a882d1b52fcd7ea6096bcb41b1ac4546ba38fc6e96c4185f70da143d4d07"} Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.842617 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.863735 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" event={"ID":"85bafa11-dcf4-4ea4-b68e-e2d4799e813d","Type":"ContainerStarted","Data":"97f4b6e13add8c0de66938ad08102ccf2744436b8b063f009992873921948449"} Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.866800 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" podStartSLOduration=2.348240314 podStartE2EDuration="43.866783592s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.783721647 +0000 UTC m=+1866.526432242" lastFinishedPulling="2025-12-03 22:13:31.302264885 +0000 UTC m=+1908.044975520" observedRunningTime="2025-12-03 22:13:31.861826179 +0000 UTC m=+1908.604536774" watchObservedRunningTime="2025-12-03 22:13:31.866783592 +0000 UTC m=+1908.609494187" Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.868920 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" event={"ID":"dbeb21cb-a705-464d-ace4-90e1ce5084a5","Type":"ContainerStarted","Data":"54e26dddab6564348896d4dfbb255f73f4420ec359aaefc5a31ebbb34b84868d"} Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.880334 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" event={"ID":"959f935f-f997-47bc-8798-d54f7b62f2c9","Type":"ContainerStarted","Data":"d29d8f5d39123b17b8cafc40f7f8c82cb099f60d0daad651ad8d1fe129fe1435"} Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.884359 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" event={"ID":"0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e","Type":"ContainerStarted","Data":"dd4e41484f3151e9cc6fbd245e205bff72e7c8a0b12c6da7a6c1e34a12f2e193"} Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.894816 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" event={"ID":"c7e152eb-ab3c-427a-a12b-70f46b881eee","Type":"ContainerStarted","Data":"24b07a098f24a83486d8f8ba294535a346d72eb7f5d685b17790dd15762e1b30"} Dec 03 22:13:31 crc kubenswrapper[4715]: I1203 22:13:31.896120 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" event={"ID":"594f9dc7-fb50-4241-bb93-90d8039339dc","Type":"ContainerStarted","Data":"52817fe321f0f50278fce71b24f9bedf6c153dbdf4d2bba1658e35a2df3facc2"} Dec 03 22:13:31 crc kubenswrapper[4715]: E1203 22:13:31.910296 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" podUID="959f935f-f997-47bc-8798-d54f7b62f2c9" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.904424 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" event={"ID":"0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e","Type":"ContainerStarted","Data":"215b96c511b403c9eabe64852eb88eabf26ecc7cd5fddde067e2159b46a57f76"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.904806 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.906309 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" event={"ID":"b8304681-abf0-4e5c-aa37-41453f6e7731","Type":"ContainerStarted","Data":"283042805c95fc4709540415611395a96f097a050553eca77aec7da4a596619f"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.906339 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" event={"ID":"b8304681-abf0-4e5c-aa37-41453f6e7731","Type":"ContainerStarted","Data":"7d286cae2206dcfc5d027441f687ad5c79756658c36e6cd8084d29110df4a50b"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.906510 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.908366 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" event={"ID":"c7e152eb-ab3c-427a-a12b-70f46b881eee","Type":"ContainerStarted","Data":"253ed140382fab778c75fb82f6112c9a202373dd66fd2a67aacc0a29025066ae"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.908479 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.911616 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" event={"ID":"f621de25-14cd-43f5-a0a8-cf0866337715","Type":"ContainerStarted","Data":"6c59623fe4aeb6eec9d2f92852b7b1825114911c6898ca3760f7c342833d6f78"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.911648 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" event={"ID":"f621de25-14cd-43f5-a0a8-cf0866337715","Type":"ContainerStarted","Data":"d48afa49d076c403c57388b32da47be3c3f5fc1b52ac74e14aea9d9b2203cea4"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.911803 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.915302 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" event={"ID":"dbeb21cb-a705-464d-ace4-90e1ce5084a5","Type":"ContainerStarted","Data":"8555b0aa4b1116c0ab32899b476866b8f84992f92becd90725738dc32ff0fda5"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.915430 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.918657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" event={"ID":"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef","Type":"ContainerStarted","Data":"f3509bcbf7aa0f0c153981d9368ed12503b5635874f7b4ee4b008e094a423bfd"} Dec 03 22:13:32 crc kubenswrapper[4715]: E1203 22:13:32.920631 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" podUID="6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.922541 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" event={"ID":"594f9dc7-fb50-4241-bb93-90d8039339dc","Type":"ContainerStarted","Data":"aeec3560841ac4af39529f2624161aea605ef3dc701bf0c6eda1ce9f2b9e5d9d"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.922620 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.925525 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" podStartSLOduration=4.414454786 podStartE2EDuration="44.925514853s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.863186131 +0000 UTC m=+1866.605896716" lastFinishedPulling="2025-12-03 22:13:30.374246148 +0000 UTC m=+1907.116956783" observedRunningTime="2025-12-03 22:13:32.921844115 +0000 UTC m=+1909.664554710" watchObservedRunningTime="2025-12-03 22:13:32.925514853 +0000 UTC m=+1909.668225448" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.931607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" event={"ID":"6925aa7e-915e-4bbf-a3e1-991ce581b49c","Type":"ContainerStarted","Data":"cc8d906401a5651b1f5d4fe3e107641be457f92647cd02b5905e1bc48b54791a"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.932124 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.936251 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" event={"ID":"85bafa11-dcf4-4ea4-b68e-e2d4799e813d","Type":"ContainerStarted","Data":"8495b196a7a7bc07b64f44f6c377da049d81c08b2e30e050d2eb65ce71a1532f"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.936935 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.938757 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" event={"ID":"a41a8afb-d557-4caa-bf3e-57a9a5115f57","Type":"ContainerStarted","Data":"1b0b9d45116173aec45c6c5e378d15fc830544218bbef42c1af8dbd79da4d7f5"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.938841 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" event={"ID":"a41a8afb-d557-4caa-bf3e-57a9a5115f57","Type":"ContainerStarted","Data":"4e6dc2f97a4b8251ddffcc2cd3c813d27be0e3c50977b96277ebd1ccf34f68b6"} Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.938926 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:32 crc kubenswrapper[4715]: I1203 22:13:32.955362 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" podStartSLOduration=3.483839482 podStartE2EDuration="44.955349761s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.830785797 +0000 UTC m=+1866.573496392" lastFinishedPulling="2025-12-03 22:13:31.302296076 +0000 UTC m=+1908.045006671" observedRunningTime="2025-12-03 22:13:32.952901195 +0000 UTC m=+1909.695611790" watchObservedRunningTime="2025-12-03 22:13:32.955349761 +0000 UTC m=+1909.698060356" Dec 03 22:13:33 crc kubenswrapper[4715]: I1203 22:13:33.013834 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" podStartSLOduration=18.81285593 podStartE2EDuration="45.013807154s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:13:05.095647539 +0000 UTC m=+1881.838358134" lastFinishedPulling="2025-12-03 22:13:31.296598723 +0000 UTC m=+1908.039309358" observedRunningTime="2025-12-03 22:13:32.987890281 +0000 UTC m=+1909.730600876" watchObservedRunningTime="2025-12-03 22:13:33.013807154 +0000 UTC m=+1909.756517749" Dec 03 22:13:33 crc kubenswrapper[4715]: I1203 22:13:33.061351 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" podStartSLOduration=3.621056919 podStartE2EDuration="45.061326955s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.861723342 +0000 UTC m=+1866.604433937" lastFinishedPulling="2025-12-03 22:13:31.301993378 +0000 UTC m=+1908.044703973" observedRunningTime="2025-12-03 22:13:33.04395015 +0000 UTC m=+1909.786660755" watchObservedRunningTime="2025-12-03 22:13:33.061326955 +0000 UTC m=+1909.804037550" Dec 03 22:13:33 crc kubenswrapper[4715]: I1203 22:13:33.088773 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" podStartSLOduration=4.598371747 podStartE2EDuration="45.088754978s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.881981796 +0000 UTC m=+1866.624692391" lastFinishedPulling="2025-12-03 22:13:30.372364997 +0000 UTC m=+1907.115075622" observedRunningTime="2025-12-03 22:13:33.077438325 +0000 UTC m=+1909.820148920" watchObservedRunningTime="2025-12-03 22:13:33.088754978 +0000 UTC m=+1909.831465573" Dec 03 22:13:33 crc kubenswrapper[4715]: I1203 22:13:33.100954 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" podStartSLOduration=3.697753279 podStartE2EDuration="45.100891033s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.889660638 +0000 UTC m=+1866.632371233" lastFinishedPulling="2025-12-03 22:13:31.292798372 +0000 UTC m=+1908.035508987" observedRunningTime="2025-12-03 22:13:33.098140949 +0000 UTC m=+1909.840851544" watchObservedRunningTime="2025-12-03 22:13:33.100891033 +0000 UTC m=+1909.843601628" Dec 03 22:13:33 crc kubenswrapper[4715]: I1203 22:13:33.127654 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" podStartSLOduration=45.127637038 podStartE2EDuration="45.127637038s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:13:33.123290282 +0000 UTC m=+1909.866000877" watchObservedRunningTime="2025-12-03 22:13:33.127637038 +0000 UTC m=+1909.870347633" Dec 03 22:13:33 crc kubenswrapper[4715]: I1203 22:13:33.154014 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" podStartSLOduration=3.785110259 podStartE2EDuration="45.154001813s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.880396034 +0000 UTC m=+1866.623106629" lastFinishedPulling="2025-12-03 22:13:31.249287588 +0000 UTC m=+1907.991998183" observedRunningTime="2025-12-03 22:13:33.147918671 +0000 UTC m=+1909.890629276" watchObservedRunningTime="2025-12-03 22:13:33.154001813 +0000 UTC m=+1909.896712408" Dec 03 22:13:33 crc kubenswrapper[4715]: E1203 22:13:33.949632 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" podUID="6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef" Dec 03 22:13:33 crc kubenswrapper[4715]: I1203 22:13:33.968951 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" podStartSLOduration=8.31658438 podStartE2EDuration="45.968923135s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.873817451 +0000 UTC m=+1866.616528046" lastFinishedPulling="2025-12-03 22:13:27.526156196 +0000 UTC m=+1904.268866801" observedRunningTime="2025-12-03 22:13:33.176846954 +0000 UTC m=+1909.919557559" watchObservedRunningTime="2025-12-03 22:13:33.968923135 +0000 UTC m=+1910.711633760" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.076749 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-swl5h"] Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.078658 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.090295 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swl5h"] Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.175748 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-catalog-content\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.175830 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-utilities\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.175898 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jgcb\" (UniqueName: \"kubernetes.io/projected/44d017f9-924b-4a67-9e52-ca46f73464bb-kube-api-access-9jgcb\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.277861 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-catalog-content\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.277942 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-utilities\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.278005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jgcb\" (UniqueName: \"kubernetes.io/projected/44d017f9-924b-4a67-9e52-ca46f73464bb-kube-api-access-9jgcb\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.282259 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-catalog-content\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.282259 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-utilities\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.313950 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jgcb\" (UniqueName: \"kubernetes.io/projected/44d017f9-924b-4a67-9e52-ca46f73464bb-kube-api-access-9jgcb\") pod \"community-operators-swl5h\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.402022 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.939063 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swl5h"] Dec 03 22:13:36 crc kubenswrapper[4715]: I1203 22:13:36.968651 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swl5h" event={"ID":"44d017f9-924b-4a67-9e52-ca46f73464bb","Type":"ContainerStarted","Data":"366492dd023287a391d72e0a3473cc86801e763de7697d819364931a3c41e1b8"} Dec 03 22:13:37 crc kubenswrapper[4715]: I1203 22:13:37.981334 4715 generic.go:334] "Generic (PLEG): container finished" podID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerID="b0d6e3ddb761aff29847431bf1bb2a0562f433d356d53dfdad2bdb4ac03b98b2" exitCode=0 Dec 03 22:13:37 crc kubenswrapper[4715]: I1203 22:13:37.981392 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swl5h" event={"ID":"44d017f9-924b-4a67-9e52-ca46f73464bb","Type":"ContainerDied","Data":"b0d6e3ddb761aff29847431bf1bb2a0562f433d356d53dfdad2bdb4ac03b98b2"} Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.277985 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hjv26"] Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.283289 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.296482 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjv26"] Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.414408 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-utilities\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.414452 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4bjr\" (UniqueName: \"kubernetes.io/projected/94605d50-b394-4dfa-8c89-e8c04dc7abb8-kube-api-access-m4bjr\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.414572 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-catalog-content\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.516833 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-catalog-content\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.516912 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-utilities\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.516977 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4bjr\" (UniqueName: \"kubernetes.io/projected/94605d50-b394-4dfa-8c89-e8c04dc7abb8-kube-api-access-m4bjr\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.517389 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-catalog-content\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.517431 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-utilities\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.539532 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4bjr\" (UniqueName: \"kubernetes.io/projected/94605d50-b394-4dfa-8c89-e8c04dc7abb8-kube-api-access-m4bjr\") pod \"redhat-marketplace-hjv26\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.610107 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.641740 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-qbp4s" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.879188 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-hj6qn" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.956486 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-8g9vt" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.960634 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-ttfdp" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.968683 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-68hlj" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.969455 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-252wh" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.969569 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-hm2q2" Dec 03 22:13:38 crc kubenswrapper[4715]: I1203 22:13:38.969716 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-9ch2g" Dec 03 22:13:39 crc kubenswrapper[4715]: I1203 22:13:39.287002 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjv26"] Dec 03 22:13:40 crc kubenswrapper[4715]: I1203 22:13:40.138603 4715 generic.go:334] "Generic (PLEG): container finished" podID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerID="68a76e8c76dc2c5ec5797f0ea73d3a5f868e185852933a6a614a1f412b3331a2" exitCode=0 Dec 03 22:13:40 crc kubenswrapper[4715]: I1203 22:13:40.138685 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swl5h" event={"ID":"44d017f9-924b-4a67-9e52-ca46f73464bb","Type":"ContainerDied","Data":"68a76e8c76dc2c5ec5797f0ea73d3a5f868e185852933a6a614a1f412b3331a2"} Dec 03 22:13:40 crc kubenswrapper[4715]: I1203 22:13:40.141272 4715 generic.go:334] "Generic (PLEG): container finished" podID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerID="bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf" exitCode=0 Dec 03 22:13:40 crc kubenswrapper[4715]: I1203 22:13:40.141324 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjv26" event={"ID":"94605d50-b394-4dfa-8c89-e8c04dc7abb8","Type":"ContainerDied","Data":"bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf"} Dec 03 22:13:40 crc kubenswrapper[4715]: I1203 22:13:40.141362 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjv26" event={"ID":"94605d50-b394-4dfa-8c89-e8c04dc7abb8","Type":"ContainerStarted","Data":"cf13e57f360105ceff66c6c4bc618b9f0428b6306cf3d429d3fce3806afe6607"} Dec 03 22:13:41 crc kubenswrapper[4715]: I1203 22:13:41.099567 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-bd9c944fb-mkhhs" Dec 03 22:13:41 crc kubenswrapper[4715]: I1203 22:13:41.153877 4715 generic.go:334] "Generic (PLEG): container finished" podID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerID="2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78" exitCode=0 Dec 03 22:13:41 crc kubenswrapper[4715]: I1203 22:13:41.154209 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjv26" event={"ID":"94605d50-b394-4dfa-8c89-e8c04dc7abb8","Type":"ContainerDied","Data":"2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78"} Dec 03 22:13:41 crc kubenswrapper[4715]: I1203 22:13:41.164102 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swl5h" event={"ID":"44d017f9-924b-4a67-9e52-ca46f73464bb","Type":"ContainerStarted","Data":"01e261b0fe35a38d0cd3924ee35e828c92a19200555596dc7c3e37a2e311a6df"} Dec 03 22:13:41 crc kubenswrapper[4715]: I1203 22:13:41.202407 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-swl5h" podStartSLOduration=2.666560945 podStartE2EDuration="5.202389096s" podCreationTimestamp="2025-12-03 22:13:36 +0000 UTC" firstStartedPulling="2025-12-03 22:13:37.985451702 +0000 UTC m=+1914.728162337" lastFinishedPulling="2025-12-03 22:13:40.521279893 +0000 UTC m=+1917.263990488" observedRunningTime="2025-12-03 22:13:41.200060614 +0000 UTC m=+1917.942771229" watchObservedRunningTime="2025-12-03 22:13:41.202389096 +0000 UTC m=+1917.945099691" Dec 03 22:13:42 crc kubenswrapper[4715]: I1203 22:13:42.176518 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjv26" event={"ID":"94605d50-b394-4dfa-8c89-e8c04dc7abb8","Type":"ContainerStarted","Data":"bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be"} Dec 03 22:13:42 crc kubenswrapper[4715]: E1203 22:13:42.636684 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" podUID="959f935f-f997-47bc-8798-d54f7b62f2c9" Dec 03 22:13:42 crc kubenswrapper[4715]: E1203 22:13:42.637000 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" podUID="1d2cfa37-465d-4b27-8a8b-90038e1f4bf1" Dec 03 22:13:42 crc kubenswrapper[4715]: I1203 22:13:42.660325 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hjv26" podStartSLOduration=3.163659451 podStartE2EDuration="4.660305182s" podCreationTimestamp="2025-12-03 22:13:38 +0000 UTC" firstStartedPulling="2025-12-03 22:13:40.142958936 +0000 UTC m=+1916.885669571" lastFinishedPulling="2025-12-03 22:13:41.639604717 +0000 UTC m=+1918.382315302" observedRunningTime="2025-12-03 22:13:42.204181365 +0000 UTC m=+1918.946891960" watchObservedRunningTime="2025-12-03 22:13:42.660305182 +0000 UTC m=+1919.403015797" Dec 03 22:13:44 crc kubenswrapper[4715]: I1203 22:13:44.482364 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp" Dec 03 22:13:46 crc kubenswrapper[4715]: I1203 22:13:46.404182 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:46 crc kubenswrapper[4715]: I1203 22:13:46.404588 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:46 crc kubenswrapper[4715]: I1203 22:13:46.483010 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:47 crc kubenswrapper[4715]: I1203 22:13:47.277155 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:47 crc kubenswrapper[4715]: I1203 22:13:47.326345 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swl5h"] Dec 03 22:13:48 crc kubenswrapper[4715]: I1203 22:13:48.610583 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:48 crc kubenswrapper[4715]: I1203 22:13:48.610967 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:48 crc kubenswrapper[4715]: I1203 22:13:48.664904 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:49 crc kubenswrapper[4715]: I1203 22:13:49.232773 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-swl5h" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerName="registry-server" containerID="cri-o://01e261b0fe35a38d0cd3924ee35e828c92a19200555596dc7c3e37a2e311a6df" gracePeriod=2 Dec 03 22:13:49 crc kubenswrapper[4715]: I1203 22:13:49.279681 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:49 crc kubenswrapper[4715]: I1203 22:13:49.667369 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjv26"] Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.254079 4715 generic.go:334] "Generic (PLEG): container finished" podID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerID="01e261b0fe35a38d0cd3924ee35e828c92a19200555596dc7c3e37a2e311a6df" exitCode=0 Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.254427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swl5h" event={"ID":"44d017f9-924b-4a67-9e52-ca46f73464bb","Type":"ContainerDied","Data":"01e261b0fe35a38d0cd3924ee35e828c92a19200555596dc7c3e37a2e311a6df"} Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.256076 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hjv26" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerName="registry-server" containerID="cri-o://bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be" gracePeriod=2 Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.256379 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" event={"ID":"6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef","Type":"ContainerStarted","Data":"855edcf4178bb3bc70ca5979a874afc9e46229e410a0916845c5aae69e47770e"} Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.257122 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.282770 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" podStartSLOduration=17.752580436 podStartE2EDuration="1m3.282754596s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:13:05.190267509 +0000 UTC m=+1881.932978104" lastFinishedPulling="2025-12-03 22:13:50.720441669 +0000 UTC m=+1927.463152264" observedRunningTime="2025-12-03 22:13:51.277949947 +0000 UTC m=+1928.020660542" watchObservedRunningTime="2025-12-03 22:13:51.282754596 +0000 UTC m=+1928.025465201" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.515512 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.614618 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-utilities\") pod \"44d017f9-924b-4a67-9e52-ca46f73464bb\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.614760 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jgcb\" (UniqueName: \"kubernetes.io/projected/44d017f9-924b-4a67-9e52-ca46f73464bb-kube-api-access-9jgcb\") pod \"44d017f9-924b-4a67-9e52-ca46f73464bb\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.614793 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-catalog-content\") pod \"44d017f9-924b-4a67-9e52-ca46f73464bb\" (UID: \"44d017f9-924b-4a67-9e52-ca46f73464bb\") " Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.615977 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-utilities" (OuterVolumeSpecName: "utilities") pod "44d017f9-924b-4a67-9e52-ca46f73464bb" (UID: "44d017f9-924b-4a67-9e52-ca46f73464bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.622660 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44d017f9-924b-4a67-9e52-ca46f73464bb-kube-api-access-9jgcb" (OuterVolumeSpecName: "kube-api-access-9jgcb") pod "44d017f9-924b-4a67-9e52-ca46f73464bb" (UID: "44d017f9-924b-4a67-9e52-ca46f73464bb"). InnerVolumeSpecName "kube-api-access-9jgcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.637198 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.673622 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44d017f9-924b-4a67-9e52-ca46f73464bb" (UID: "44d017f9-924b-4a67-9e52-ca46f73464bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.716046 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-utilities\") pod \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.716197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-catalog-content\") pod \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.716274 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4bjr\" (UniqueName: \"kubernetes.io/projected/94605d50-b394-4dfa-8c89-e8c04dc7abb8-kube-api-access-m4bjr\") pod \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\" (UID: \"94605d50-b394-4dfa-8c89-e8c04dc7abb8\") " Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.716696 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jgcb\" (UniqueName: \"kubernetes.io/projected/44d017f9-924b-4a67-9e52-ca46f73464bb-kube-api-access-9jgcb\") on node \"crc\" DevicePath \"\"" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.716721 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.716739 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44d017f9-924b-4a67-9e52-ca46f73464bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.717999 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-utilities" (OuterVolumeSpecName: "utilities") pod "94605d50-b394-4dfa-8c89-e8c04dc7abb8" (UID: "94605d50-b394-4dfa-8c89-e8c04dc7abb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.719685 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94605d50-b394-4dfa-8c89-e8c04dc7abb8-kube-api-access-m4bjr" (OuterVolumeSpecName: "kube-api-access-m4bjr") pod "94605d50-b394-4dfa-8c89-e8c04dc7abb8" (UID: "94605d50-b394-4dfa-8c89-e8c04dc7abb8"). InnerVolumeSpecName "kube-api-access-m4bjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.743457 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94605d50-b394-4dfa-8c89-e8c04dc7abb8" (UID: "94605d50-b394-4dfa-8c89-e8c04dc7abb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.818030 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.818078 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94605d50-b394-4dfa-8c89-e8c04dc7abb8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:13:51 crc kubenswrapper[4715]: I1203 22:13:51.818099 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4bjr\" (UniqueName: \"kubernetes.io/projected/94605d50-b394-4dfa-8c89-e8c04dc7abb8-kube-api-access-m4bjr\") on node \"crc\" DevicePath \"\"" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.267822 4715 generic.go:334] "Generic (PLEG): container finished" podID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerID="bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be" exitCode=0 Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.267892 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjv26" event={"ID":"94605d50-b394-4dfa-8c89-e8c04dc7abb8","Type":"ContainerDied","Data":"bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be"} Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.267921 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hjv26" event={"ID":"94605d50-b394-4dfa-8c89-e8c04dc7abb8","Type":"ContainerDied","Data":"cf13e57f360105ceff66c6c4bc618b9f0428b6306cf3d429d3fce3806afe6607"} Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.267960 4715 scope.go:117] "RemoveContainer" containerID="bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.267967 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hjv26" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.271593 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swl5h" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.271625 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swl5h" event={"ID":"44d017f9-924b-4a67-9e52-ca46f73464bb","Type":"ContainerDied","Data":"366492dd023287a391d72e0a3473cc86801e763de7697d819364931a3c41e1b8"} Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.293476 4715 scope.go:117] "RemoveContainer" containerID="2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.318770 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjv26"] Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.323003 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hjv26"] Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.339173 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swl5h"] Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.341701 4715 scope.go:117] "RemoveContainer" containerID="bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.343876 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-swl5h"] Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.369186 4715 scope.go:117] "RemoveContainer" containerID="bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be" Dec 03 22:13:52 crc kubenswrapper[4715]: E1203 22:13:52.369987 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be\": container with ID starting with bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be not found: ID does not exist" containerID="bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.370041 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be"} err="failed to get container status \"bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be\": rpc error: code = NotFound desc = could not find container \"bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be\": container with ID starting with bec68426374b5c6ef8c69228ecb6fc59dfc180bf5311a1305301ee17c19b83be not found: ID does not exist" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.370075 4715 scope.go:117] "RemoveContainer" containerID="2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78" Dec 03 22:13:52 crc kubenswrapper[4715]: E1203 22:13:52.370493 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78\": container with ID starting with 2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78 not found: ID does not exist" containerID="2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.370611 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78"} err="failed to get container status \"2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78\": rpc error: code = NotFound desc = could not find container \"2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78\": container with ID starting with 2e03f83065ab23259bc6b2864feda8b7c1206ee8685c36649456fa6628420e78 not found: ID does not exist" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.370636 4715 scope.go:117] "RemoveContainer" containerID="bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf" Dec 03 22:13:52 crc kubenswrapper[4715]: E1203 22:13:52.371278 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf\": container with ID starting with bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf not found: ID does not exist" containerID="bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.371320 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf"} err="failed to get container status \"bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf\": rpc error: code = NotFound desc = could not find container \"bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf\": container with ID starting with bcb2f012b9c340fdd3142f2fe64584e28730c25cb514679b5cd69b32858374cf not found: ID does not exist" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.371334 4715 scope.go:117] "RemoveContainer" containerID="01e261b0fe35a38d0cd3924ee35e828c92a19200555596dc7c3e37a2e311a6df" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.395169 4715 scope.go:117] "RemoveContainer" containerID="68a76e8c76dc2c5ec5797f0ea73d3a5f868e185852933a6a614a1f412b3331a2" Dec 03 22:13:52 crc kubenswrapper[4715]: I1203 22:13:52.418880 4715 scope.go:117] "RemoveContainer" containerID="b0d6e3ddb761aff29847431bf1bb2a0562f433d356d53dfdad2bdb4ac03b98b2" Dec 03 22:13:53 crc kubenswrapper[4715]: I1203 22:13:53.644325 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" path="/var/lib/kubelet/pods/44d017f9-924b-4a67-9e52-ca46f73464bb/volumes" Dec 03 22:13:53 crc kubenswrapper[4715]: I1203 22:13:53.645670 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" path="/var/lib/kubelet/pods/94605d50-b394-4dfa-8c89-e8c04dc7abb8/volumes" Dec 03 22:13:58 crc kubenswrapper[4715]: I1203 22:13:58.343054 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" event={"ID":"1d2cfa37-465d-4b27-8a8b-90038e1f4bf1","Type":"ContainerStarted","Data":"5c568ce5726adae4158e6725bd3b1be0ea6298ec17f290f20d0d61d492bba8e5"} Dec 03 22:13:58 crc kubenswrapper[4715]: I1203 22:13:58.345243 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" event={"ID":"959f935f-f997-47bc-8798-d54f7b62f2c9","Type":"ContainerStarted","Data":"587e24449b1048531f033469c5c5e6aaaf5b981cc6f5c94ee097748e60bb6dc4"} Dec 03 22:13:58 crc kubenswrapper[4715]: I1203 22:13:58.345463 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" Dec 03 22:13:58 crc kubenswrapper[4715]: I1203 22:13:58.366546 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-hnzzx" podStartSLOduration=3.077721994 podStartE2EDuration="1m10.366523253s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.900814382 +0000 UTC m=+1866.643524977" lastFinishedPulling="2025-12-03 22:13:57.189615641 +0000 UTC m=+1933.932326236" observedRunningTime="2025-12-03 22:13:58.363199935 +0000 UTC m=+1935.105910540" watchObservedRunningTime="2025-12-03 22:13:58.366523253 +0000 UTC m=+1935.109233858" Dec 03 22:13:58 crc kubenswrapper[4715]: I1203 22:13:58.384350 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" podStartSLOduration=3.036385387 podStartE2EDuration="1m10.38433455s" podCreationTimestamp="2025-12-03 22:12:48 +0000 UTC" firstStartedPulling="2025-12-03 22:12:49.880762934 +0000 UTC m=+1866.623473529" lastFinishedPulling="2025-12-03 22:13:57.228712087 +0000 UTC m=+1933.971422692" observedRunningTime="2025-12-03 22:13:58.382565782 +0000 UTC m=+1935.125276387" watchObservedRunningTime="2025-12-03 22:13:58.38433455 +0000 UTC m=+1935.127045145" Dec 03 22:14:04 crc kubenswrapper[4715]: I1203 22:14:04.476328 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-kwnlp" Dec 03 22:14:08 crc kubenswrapper[4715]: I1203 22:14:08.978099 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-gtpjt" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.836753 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8h7l5"] Dec 03 22:14:24 crc kubenswrapper[4715]: E1203 22:14:24.837877 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerName="extract-utilities" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.837893 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerName="extract-utilities" Dec 03 22:14:24 crc kubenswrapper[4715]: E1203 22:14:24.837913 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerName="registry-server" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.837922 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerName="registry-server" Dec 03 22:14:24 crc kubenswrapper[4715]: E1203 22:14:24.837949 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerName="extract-content" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.837956 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerName="extract-content" Dec 03 22:14:24 crc kubenswrapper[4715]: E1203 22:14:24.837979 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerName="extract-utilities" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.837987 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerName="extract-utilities" Dec 03 22:14:24 crc kubenswrapper[4715]: E1203 22:14:24.838005 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerName="extract-content" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.838012 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerName="extract-content" Dec 03 22:14:24 crc kubenswrapper[4715]: E1203 22:14:24.838025 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerName="registry-server" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.838033 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerName="registry-server" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.838197 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="94605d50-b394-4dfa-8c89-e8c04dc7abb8" containerName="registry-server" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.838225 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="44d017f9-924b-4a67-9e52-ca46f73464bb" containerName="registry-server" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.839121 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.842123 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.842705 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-4j5k2" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.843077 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.843354 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.843630 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.852263 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8h7l5"] Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.919561 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5fzt\" (UniqueName: \"kubernetes.io/projected/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-kube-api-access-z5fzt\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.919638 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-config\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:24 crc kubenswrapper[4715]: I1203 22:14:24.919699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:25 crc kubenswrapper[4715]: I1203 22:14:25.020771 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-config\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:25 crc kubenswrapper[4715]: I1203 22:14:25.020844 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:25 crc kubenswrapper[4715]: I1203 22:14:25.020879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5fzt\" (UniqueName: \"kubernetes.io/projected/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-kube-api-access-z5fzt\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:25 crc kubenswrapper[4715]: I1203 22:14:25.021935 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-config\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:25 crc kubenswrapper[4715]: I1203 22:14:25.021950 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:25 crc kubenswrapper[4715]: I1203 22:14:25.049764 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5fzt\" (UniqueName: \"kubernetes.io/projected/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-kube-api-access-z5fzt\") pod \"dnsmasq-dns-78dd6ddcc-8h7l5\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:25 crc kubenswrapper[4715]: I1203 22:14:25.166239 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:25 crc kubenswrapper[4715]: I1203 22:14:25.666121 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8h7l5"] Dec 03 22:14:26 crc kubenswrapper[4715]: I1203 22:14:26.628881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" event={"ID":"95f54df8-0793-4e50-b7bd-fd3d2745b9a0","Type":"ContainerStarted","Data":"09347b40c819100e219fa0d06421409f8fbd3cf179b8f55375f58a762d689b57"} Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.845480 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-94s7g"] Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.847191 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.856676 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-dns-svc\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.856788 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-config\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.856946 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmdvs\" (UniqueName: \"kubernetes.io/projected/921209ad-aef3-4407-9bde-3ab12ea68e43-kube-api-access-mmdvs\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.873930 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-94s7g"] Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.959364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-dns-svc\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.959429 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-config\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.959478 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmdvs\" (UniqueName: \"kubernetes.io/projected/921209ad-aef3-4407-9bde-3ab12ea68e43-kube-api-access-mmdvs\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.960296 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-dns-svc\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.960302 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-config\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:27 crc kubenswrapper[4715]: I1203 22:14:27.978264 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmdvs\" (UniqueName: \"kubernetes.io/projected/921209ad-aef3-4407-9bde-3ab12ea68e43-kube-api-access-mmdvs\") pod \"dnsmasq-dns-666b6646f7-94s7g\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.165930 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.251086 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8h7l5"] Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.289916 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vq29"] Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.291446 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.306323 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vq29"] Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.363456 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg7nn\" (UniqueName: \"kubernetes.io/projected/28d0a607-f5ba-4e00-9e3d-111f0132026e-kube-api-access-bg7nn\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.363563 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-config\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.363638 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.464923 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-config\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.465005 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.465028 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg7nn\" (UniqueName: \"kubernetes.io/projected/28d0a607-f5ba-4e00-9e3d-111f0132026e-kube-api-access-bg7nn\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.466112 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.466133 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-config\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.482691 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg7nn\" (UniqueName: \"kubernetes.io/projected/28d0a607-f5ba-4e00-9e3d-111f0132026e-kube-api-access-bg7nn\") pod \"dnsmasq-dns-57d769cc4f-5vq29\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:28 crc kubenswrapper[4715]: I1203 22:14:28.641076 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.026530 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.027612 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.031594 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.031717 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7pp49" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.032010 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.032067 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.032152 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.032349 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.032592 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.040833 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.174653 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.174699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.174808 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.174916 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-config-data\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.174982 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.175047 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.175138 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.175186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.175226 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.175329 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.175367 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-894b7\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-kube-api-access-894b7\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.276890 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.277310 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.277340 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.277364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.277409 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.277406 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.277435 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.278083 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-894b7\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-kube-api-access-894b7\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.278221 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.278263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.278285 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.278343 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-config-data\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.279069 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.279137 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.279267 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-config-data\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.279621 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.280140 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.281211 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.282956 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.283576 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.293296 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.297202 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-894b7\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-kube-api-access-894b7\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.302557 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.361267 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.411348 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.413403 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.415681 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.417935 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.418044 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.418071 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.418078 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.418867 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-8mhj2" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.419603 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.421010 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d6205840-7ea5-4060-844a-f6ce9feed200-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582223 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmqlc\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-kube-api-access-cmqlc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582269 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582289 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582309 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582326 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582341 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582356 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582567 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.582655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d6205840-7ea5-4060-844a-f6ce9feed200-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.684599 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.684679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d6205840-7ea5-4060-844a-f6ce9feed200-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.684803 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d6205840-7ea5-4060-844a-f6ce9feed200-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.684863 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.684902 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmqlc\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-kube-api-access-cmqlc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.685064 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.685400 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.689177 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.689236 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.689404 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.689627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.689666 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.690733 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.690841 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.691029 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d6205840-7ea5-4060-844a-f6ce9feed200-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.694242 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.694670 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.696883 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d6205840-7ea5-4060-844a-f6ce9feed200-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.697847 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.699187 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.699922 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmqlc\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-kube-api-access-cmqlc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.701290 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.728913 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:29 crc kubenswrapper[4715]: I1203 22:14:29.745753 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:14:30 crc kubenswrapper[4715]: I1203 22:14:30.854003 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 22:14:30 crc kubenswrapper[4715]: I1203 22:14:30.857415 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 22:14:30 crc kubenswrapper[4715]: I1203 22:14:30.863982 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 22:14:30 crc kubenswrapper[4715]: I1203 22:14:30.870487 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 22:14:30 crc kubenswrapper[4715]: I1203 22:14:30.870851 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 22:14:30 crc kubenswrapper[4715]: I1203 22:14:30.872029 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 22:14:30 crc kubenswrapper[4715]: I1203 22:14:30.872481 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-88lq5" Dec 03 22:14:30 crc kubenswrapper[4715]: I1203 22:14:30.882166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.009450 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-kolla-config\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.009531 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.009582 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b024438-6200-45dd-9297-ae2b071dae15-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.009633 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b024438-6200-45dd-9297-ae2b071dae15-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.009655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-config-data-default\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.009686 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tmd7\" (UniqueName: \"kubernetes.io/projected/3b024438-6200-45dd-9297-ae2b071dae15-kube-api-access-4tmd7\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.009732 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.009759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b024438-6200-45dd-9297-ae2b071dae15-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111556 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b024438-6200-45dd-9297-ae2b071dae15-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111606 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-config-data-default\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111646 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tmd7\" (UniqueName: \"kubernetes.io/projected/3b024438-6200-45dd-9297-ae2b071dae15-kube-api-access-4tmd7\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111697 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111743 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b024438-6200-45dd-9297-ae2b071dae15-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111778 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-kolla-config\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111805 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111850 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b024438-6200-45dd-9297-ae2b071dae15-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.111894 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.117911 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3b024438-6200-45dd-9297-ae2b071dae15-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.125378 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-config-data-default\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.127752 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3b024438-6200-45dd-9297-ae2b071dae15-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.128473 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-kolla-config\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.128819 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3b024438-6200-45dd-9297-ae2b071dae15-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.131071 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b024438-6200-45dd-9297-ae2b071dae15-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.135057 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tmd7\" (UniqueName: \"kubernetes.io/projected/3b024438-6200-45dd-9297-ae2b071dae15-kube-api-access-4tmd7\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.158688 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"3b024438-6200-45dd-9297-ae2b071dae15\") " pod="openstack/openstack-galera-0" Dec 03 22:14:31 crc kubenswrapper[4715]: I1203 22:14:31.224119 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.430110 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.431635 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.435190 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.435256 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-jwz8d" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.435710 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.435837 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.448417 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.533937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.533997 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jvp8\" (UniqueName: \"kubernetes.io/projected/b487c9d6-b74d-4db6-8ef4-caee02c83106-kube-api-access-2jvp8\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.534025 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b487c9d6-b74d-4db6-8ef4-caee02c83106-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.534056 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.534076 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.534095 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b487c9d6-b74d-4db6-8ef4-caee02c83106-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.534135 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b487c9d6-b74d-4db6-8ef4-caee02c83106-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.534179 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.635422 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b487c9d6-b74d-4db6-8ef4-caee02c83106-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.635489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.635542 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.635568 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jvp8\" (UniqueName: \"kubernetes.io/projected/b487c9d6-b74d-4db6-8ef4-caee02c83106-kube-api-access-2jvp8\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.635590 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b487c9d6-b74d-4db6-8ef4-caee02c83106-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.635616 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.635634 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.635649 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b487c9d6-b74d-4db6-8ef4-caee02c83106-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.636664 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.636898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b487c9d6-b74d-4db6-8ef4-caee02c83106-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.637082 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.637717 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.638270 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b487c9d6-b74d-4db6-8ef4-caee02c83106-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.639677 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b487c9d6-b74d-4db6-8ef4-caee02c83106-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.644999 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b487c9d6-b74d-4db6-8ef4-caee02c83106-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.657599 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jvp8\" (UniqueName: \"kubernetes.io/projected/b487c9d6-b74d-4db6-8ef4-caee02c83106-kube-api-access-2jvp8\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.673931 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"b487c9d6-b74d-4db6-8ef4-caee02c83106\") " pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.685787 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.686859 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.699237 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.701904 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.702241 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-smhqr" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.702628 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.740315 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.740379 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.740408 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-config-data\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.740431 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqzm4\" (UniqueName: \"kubernetes.io/projected/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-kube-api-access-zqzm4\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.740461 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-kolla-config\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.755433 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.842090 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.842134 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.842153 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-config-data\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.842172 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqzm4\" (UniqueName: \"kubernetes.io/projected/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-kube-api-access-zqzm4\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.842200 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-kolla-config\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.842844 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-config-data\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.843062 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-kolla-config\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.849770 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.860324 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:32 crc kubenswrapper[4715]: I1203 22:14:32.863028 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqzm4\" (UniqueName: \"kubernetes.io/projected/5ffb2b06-021a-49c0-abfe-6bb5c8acba3d-kube-api-access-zqzm4\") pod \"memcached-0\" (UID: \"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d\") " pod="openstack/memcached-0" Dec 03 22:14:33 crc kubenswrapper[4715]: I1203 22:14:33.060428 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 22:14:34 crc kubenswrapper[4715]: I1203 22:14:34.650384 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:14:34 crc kubenswrapper[4715]: I1203 22:14:34.654877 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 22:14:34 crc kubenswrapper[4715]: I1203 22:14:34.657492 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-spqgk" Dec 03 22:14:34 crc kubenswrapper[4715]: I1203 22:14:34.663284 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:14:34 crc kubenswrapper[4715]: I1203 22:14:34.773310 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnp5f\" (UniqueName: \"kubernetes.io/projected/e514cbad-5355-479e-be39-a625b874551c-kube-api-access-fnp5f\") pod \"kube-state-metrics-0\" (UID: \"e514cbad-5355-479e-be39-a625b874551c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:14:34 crc kubenswrapper[4715]: I1203 22:14:34.874998 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnp5f\" (UniqueName: \"kubernetes.io/projected/e514cbad-5355-479e-be39-a625b874551c-kube-api-access-fnp5f\") pod \"kube-state-metrics-0\" (UID: \"e514cbad-5355-479e-be39-a625b874551c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:14:34 crc kubenswrapper[4715]: I1203 22:14:34.899537 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnp5f\" (UniqueName: \"kubernetes.io/projected/e514cbad-5355-479e-be39-a625b874551c-kube-api-access-fnp5f\") pod \"kube-state-metrics-0\" (UID: \"e514cbad-5355-479e-be39-a625b874551c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:14:35 crc kubenswrapper[4715]: I1203 22:14:35.015790 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.585227 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5fzvj"] Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.586812 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.588744 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-65ht4" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.589159 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.589401 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.601363 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-xq7sg"] Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.603391 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.616852 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5fzvj"] Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.616925 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xq7sg"] Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654052 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/893e6b97-27e2-4372-8fc6-90562f2781bb-ovn-controller-tls-certs\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654104 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-lib\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654380 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-run\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654448 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-run-ovn\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654471 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/893e6b97-27e2-4372-8fc6-90562f2781bb-scripts\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654561 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-etc-ovs\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654714 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-scripts\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654901 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893e6b97-27e2-4372-8fc6-90562f2781bb-combined-ca-bundle\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.654946 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-run\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.655008 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-log-ovn\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.655043 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5zd7\" (UniqueName: \"kubernetes.io/projected/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-kube-api-access-w5zd7\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.655075 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-log\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.655094 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvjwb\" (UniqueName: \"kubernetes.io/projected/893e6b97-27e2-4372-8fc6-90562f2781bb-kube-api-access-lvjwb\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.757995 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893e6b97-27e2-4372-8fc6-90562f2781bb-combined-ca-bundle\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758055 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-run\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758101 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-log-ovn\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758127 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5zd7\" (UniqueName: \"kubernetes.io/projected/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-kube-api-access-w5zd7\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758157 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-log\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758198 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvjwb\" (UniqueName: \"kubernetes.io/projected/893e6b97-27e2-4372-8fc6-90562f2781bb-kube-api-access-lvjwb\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758231 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/893e6b97-27e2-4372-8fc6-90562f2781bb-ovn-controller-tls-certs\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758261 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-lib\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-run\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758344 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-run-ovn\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758367 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/893e6b97-27e2-4372-8fc6-90562f2781bb-scripts\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758392 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-etc-ovs\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.758429 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-scripts\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.763341 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-run\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.763341 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-run\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.763389 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-log-ovn\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.763412 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-log\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.763434 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/893e6b97-27e2-4372-8fc6-90562f2781bb-var-run-ovn\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.763783 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-etc-ovs\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.763837 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-var-lib\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.770544 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/893e6b97-27e2-4372-8fc6-90562f2781bb-combined-ca-bundle\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.772180 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/893e6b97-27e2-4372-8fc6-90562f2781bb-scripts\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.772621 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-scripts\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.778698 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/893e6b97-27e2-4372-8fc6-90562f2781bb-ovn-controller-tls-certs\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.786694 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5zd7\" (UniqueName: \"kubernetes.io/projected/d495583b-de78-4dd5-aa7b-70e15ee3cf0e-kube-api-access-w5zd7\") pod \"ovn-controller-ovs-xq7sg\" (UID: \"d495583b-de78-4dd5-aa7b-70e15ee3cf0e\") " pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.788272 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvjwb\" (UniqueName: \"kubernetes.io/projected/893e6b97-27e2-4372-8fc6-90562f2781bb-kube-api-access-lvjwb\") pod \"ovn-controller-5fzvj\" (UID: \"893e6b97-27e2-4372-8fc6-90562f2781bb\") " pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.960112 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:38 crc kubenswrapper[4715]: I1203 22:14:38.986923 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.514153 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.516473 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.522784 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.522999 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.523177 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.523401 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-r94vh" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.523590 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.535991 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.591085 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.591166 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af93de35-2858-4a66-80a3-b7be3a28ff3f-config\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.591278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.591413 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af93de35-2858-4a66-80a3-b7be3a28ff3f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.591665 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4fzl\" (UniqueName: \"kubernetes.io/projected/af93de35-2858-4a66-80a3-b7be3a28ff3f-kube-api-access-d4fzl\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.591864 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.592062 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/af93de35-2858-4a66-80a3-b7be3a28ff3f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.592122 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.693582 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af93de35-2858-4a66-80a3-b7be3a28ff3f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.693651 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4fzl\" (UniqueName: \"kubernetes.io/projected/af93de35-2858-4a66-80a3-b7be3a28ff3f-kube-api-access-d4fzl\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.693715 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.693778 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/af93de35-2858-4a66-80a3-b7be3a28ff3f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.693798 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.693830 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.693859 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af93de35-2858-4a66-80a3-b7be3a28ff3f-config\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.693904 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.694357 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.694592 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/af93de35-2858-4a66-80a3-b7be3a28ff3f-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.695203 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af93de35-2858-4a66-80a3-b7be3a28ff3f-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.695218 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af93de35-2858-4a66-80a3-b7be3a28ff3f-config\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.700277 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.700736 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.704222 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/af93de35-2858-4a66-80a3-b7be3a28ff3f-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.721919 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4fzl\" (UniqueName: \"kubernetes.io/projected/af93de35-2858-4a66-80a3-b7be3a28ff3f-kube-api-access-d4fzl\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.725784 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"af93de35-2858-4a66-80a3-b7be3a28ff3f\") " pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:40 crc kubenswrapper[4715]: I1203 22:14:40.850723 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.237244 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.239049 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.243138 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.243360 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.243734 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-c62bl" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.243960 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.246202 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.320102 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/695ea890-c971-4f5a-84ca-fbd83fa5d74a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.320193 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.320306 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.320379 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.320586 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/695ea890-c971-4f5a-84ca-fbd83fa5d74a-config\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.320628 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/695ea890-c971-4f5a-84ca-fbd83fa5d74a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.320671 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.320735 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvjpm\" (UniqueName: \"kubernetes.io/projected/695ea890-c971-4f5a-84ca-fbd83fa5d74a-kube-api-access-rvjpm\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.422666 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/695ea890-c971-4f5a-84ca-fbd83fa5d74a-config\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.422779 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/695ea890-c971-4f5a-84ca-fbd83fa5d74a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.422838 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.422889 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvjpm\" (UniqueName: \"kubernetes.io/projected/695ea890-c971-4f5a-84ca-fbd83fa5d74a-kube-api-access-rvjpm\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.422991 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/695ea890-c971-4f5a-84ca-fbd83fa5d74a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.423066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.423107 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.423154 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.424168 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.424235 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/695ea890-c971-4f5a-84ca-fbd83fa5d74a-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.425602 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/695ea890-c971-4f5a-84ca-fbd83fa5d74a-config\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.425988 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/695ea890-c971-4f5a-84ca-fbd83fa5d74a-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.429326 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.434781 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.440988 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/695ea890-c971-4f5a-84ca-fbd83fa5d74a-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.460705 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvjpm\" (UniqueName: \"kubernetes.io/projected/695ea890-c971-4f5a-84ca-fbd83fa5d74a-kube-api-access-rvjpm\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.470443 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"695ea890-c971-4f5a-84ca-fbd83fa5d74a\") " pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:42 crc kubenswrapper[4715]: I1203 22:14:42.579233 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:43 crc kubenswrapper[4715]: I1203 22:14:43.152092 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:14:43 crc kubenswrapper[4715]: W1203 22:14:43.624358 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e2a0d28_27e4_4a86_8bf7_d5007af5162c.slice/crio-81d7bb66259f93ed78635e199b00099870547e3e598c3da06b3571c490a8b28c WatchSource:0}: Error finding container 81d7bb66259f93ed78635e199b00099870547e3e598c3da06b3571c490a8b28c: Status 404 returned error can't find the container with id 81d7bb66259f93ed78635e199b00099870547e3e598c3da06b3571c490a8b28c Dec 03 22:14:43 crc kubenswrapper[4715]: E1203 22:14:43.624473 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 22:14:43 crc kubenswrapper[4715]: E1203 22:14:43.625019 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5fzt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-8h7l5_openstack(95f54df8-0793-4e50-b7bd-fd3d2745b9a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:14:43 crc kubenswrapper[4715]: E1203 22:14:43.626267 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" podUID="95f54df8-0793-4e50-b7bd-fd3d2745b9a0" Dec 03 22:14:43 crc kubenswrapper[4715]: I1203 22:14:43.791316 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5e2a0d28-27e4-4a86-8bf7-d5007af5162c","Type":"ContainerStarted","Data":"81d7bb66259f93ed78635e199b00099870547e3e598c3da06b3571c490a8b28c"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.179907 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:44 crc kubenswrapper[4715]: W1203 22:14:44.238105 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6205840_7ea5_4060_844a_f6ce9feed200.slice/crio-c2c285d63f37c200c967db8a0bdf6f3546f1f9cfa79a99c27918b3f77ad7aad9 WatchSource:0}: Error finding container c2c285d63f37c200c967db8a0bdf6f3546f1f9cfa79a99c27918b3f77ad7aad9: Status 404 returned error can't find the container with id c2c285d63f37c200c967db8a0bdf6f3546f1f9cfa79a99c27918b3f77ad7aad9 Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.243331 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.257826 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-config\") pod \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.257933 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-dns-svc\") pod \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.258791 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5fzt\" (UniqueName: \"kubernetes.io/projected/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-kube-api-access-z5fzt\") pod \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\" (UID: \"95f54df8-0793-4e50-b7bd-fd3d2745b9a0\") " Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.259721 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "95f54df8-0793-4e50-b7bd-fd3d2745b9a0" (UID: "95f54df8-0793-4e50-b7bd-fd3d2745b9a0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.260417 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-config" (OuterVolumeSpecName: "config") pod "95f54df8-0793-4e50-b7bd-fd3d2745b9a0" (UID: "95f54df8-0793-4e50-b7bd-fd3d2745b9a0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.265645 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.268131 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-kube-api-access-z5fzt" (OuterVolumeSpecName: "kube-api-access-z5fzt") pod "95f54df8-0793-4e50-b7bd-fd3d2745b9a0" (UID: "95f54df8-0793-4e50-b7bd-fd3d2745b9a0"). InnerVolumeSpecName "kube-api-access-z5fzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.360672 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.361599 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.361619 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.361629 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5fzt\" (UniqueName: \"kubernetes.io/projected/95f54df8-0793-4e50-b7bd-fd3d2745b9a0-kube-api-access-z5fzt\") on node \"crc\" DevicePath \"\"" Dec 03 22:14:44 crc kubenswrapper[4715]: W1203 22:14:44.364808 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ffb2b06_021a_49c0_abfe_6bb5c8acba3d.slice/crio-ddadbb8876bbc3109dbf94ac4c17b6dbe23bee309df4ab16ad656a987cb75141 WatchSource:0}: Error finding container ddadbb8876bbc3109dbf94ac4c17b6dbe23bee309df4ab16ad656a987cb75141: Status 404 returned error can't find the container with id ddadbb8876bbc3109dbf94ac4c17b6dbe23bee309df4ab16ad656a987cb75141 Dec 03 22:14:44 crc kubenswrapper[4715]: W1203 22:14:44.367677 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb487c9d6_b74d_4db6_8ef4_caee02c83106.slice/crio-a29f354f7c542c9a9cd2f48bd5f77d8bf30d9955002b5676f8e16f828c831c54 WatchSource:0}: Error finding container a29f354f7c542c9a9cd2f48bd5f77d8bf30d9955002b5676f8e16f828c831c54: Status 404 returned error can't find the container with id a29f354f7c542c9a9cd2f48bd5f77d8bf30d9955002b5676f8e16f828c831c54 Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.369389 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.464363 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xq7sg"] Dec 03 22:14:44 crc kubenswrapper[4715]: W1203 22:14:44.466198 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd495583b_de78_4dd5_aa7b_70e15ee3cf0e.slice/crio-ac3be04eaad203c9739efe806f2c3369bd6c887333cac4b9028a7ea55594026f WatchSource:0}: Error finding container ac3be04eaad203c9739efe806f2c3369bd6c887333cac4b9028a7ea55594026f: Status 404 returned error can't find the container with id ac3be04eaad203c9739efe806f2c3369bd6c887333cac4b9028a7ea55594026f Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.521858 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5fzvj"] Dec 03 22:14:44 crc kubenswrapper[4715]: W1203 22:14:44.523028 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod893e6b97_27e2_4372_8fc6_90562f2781bb.slice/crio-5e79d4045eae593fca685f8a567584095ba9861cbd82e3a3118360bc9e552bab WatchSource:0}: Error finding container 5e79d4045eae593fca685f8a567584095ba9861cbd82e3a3118360bc9e552bab: Status 404 returned error can't find the container with id 5e79d4045eae593fca685f8a567584095ba9861cbd82e3a3118360bc9e552bab Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.533227 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.541589 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-94s7g"] Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.548197 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vq29"] Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.595123 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 22:14:44 crc kubenswrapper[4715]: W1203 22:14:44.609057 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod695ea890_c971_4f5a_84ca_fbd83fa5d74a.slice/crio-d1bba5c0e8ac5c7972ba945d3ab7e03e413829dbdae23fc0f6e164d7c4f7602b WatchSource:0}: Error finding container d1bba5c0e8ac5c7972ba945d3ab7e03e413829dbdae23fc0f6e164d7c4f7602b: Status 404 returned error can't find the container with id d1bba5c0e8ac5c7972ba945d3ab7e03e413829dbdae23fc0f6e164d7c4f7602b Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.798053 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" event={"ID":"28d0a607-f5ba-4e00-9e3d-111f0132026e","Type":"ContainerStarted","Data":"008cdfd19e1ed9d5285c7b0971e7e1aea81047736d747d70f601e79971cae7bf"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.799311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d","Type":"ContainerStarted","Data":"ddadbb8876bbc3109dbf94ac4c17b6dbe23bee309df4ab16ad656a987cb75141"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.800459 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"b487c9d6-b74d-4db6-8ef4-caee02c83106","Type":"ContainerStarted","Data":"a29f354f7c542c9a9cd2f48bd5f77d8bf30d9955002b5676f8e16f828c831c54"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.801523 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3b024438-6200-45dd-9297-ae2b071dae15","Type":"ContainerStarted","Data":"98a8be626d07ea1169cf9b9dcdee2273320301488bbf9d9a453ca11623ef7328"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.802879 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5fzvj" event={"ID":"893e6b97-27e2-4372-8fc6-90562f2781bb","Type":"ContainerStarted","Data":"5e79d4045eae593fca685f8a567584095ba9861cbd82e3a3118360bc9e552bab"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.804387 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"695ea890-c971-4f5a-84ca-fbd83fa5d74a","Type":"ContainerStarted","Data":"d1bba5c0e8ac5c7972ba945d3ab7e03e413829dbdae23fc0f6e164d7c4f7602b"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.805607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xq7sg" event={"ID":"d495583b-de78-4dd5-aa7b-70e15ee3cf0e","Type":"ContainerStarted","Data":"ac3be04eaad203c9739efe806f2c3369bd6c887333cac4b9028a7ea55594026f"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.807148 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" event={"ID":"921209ad-aef3-4407-9bde-3ab12ea68e43","Type":"ContainerStarted","Data":"b2bf99d7bd896eab645122c28bbd3e83fba5a7d2de8d938c5c949769a06f4664"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.808527 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d6205840-7ea5-4060-844a-f6ce9feed200","Type":"ContainerStarted","Data":"c2c285d63f37c200c967db8a0bdf6f3546f1f9cfa79a99c27918b3f77ad7aad9"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.809728 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" event={"ID":"95f54df8-0793-4e50-b7bd-fd3d2745b9a0","Type":"ContainerDied","Data":"09347b40c819100e219fa0d06421409f8fbd3cf179b8f55375f58a762d689b57"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.809749 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8h7l5" Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.810909 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e514cbad-5355-479e-be39-a625b874551c","Type":"ContainerStarted","Data":"5f134ca1511de95331999f4b8b6411c777749f153cf5ef534a6c792cac9eeadf"} Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.884224 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8h7l5"] Dec 03 22:14:44 crc kubenswrapper[4715]: I1203 22:14:44.887135 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8h7l5"] Dec 03 22:14:45 crc kubenswrapper[4715]: I1203 22:14:45.325133 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 22:14:45 crc kubenswrapper[4715]: I1203 22:14:45.643464 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95f54df8-0793-4e50-b7bd-fd3d2745b9a0" path="/var/lib/kubelet/pods/95f54df8-0793-4e50-b7bd-fd3d2745b9a0/volumes" Dec 03 22:14:45 crc kubenswrapper[4715]: I1203 22:14:45.820761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"af93de35-2858-4a66-80a3-b7be3a28ff3f","Type":"ContainerStarted","Data":"19dc154b3bcfedc964c22c2d8d3bcd41227d9fb2c93d723fca32b4beb955fd2a"} Dec 03 22:14:45 crc kubenswrapper[4715]: I1203 22:14:45.822861 4715 generic.go:334] "Generic (PLEG): container finished" podID="921209ad-aef3-4407-9bde-3ab12ea68e43" containerID="0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e" exitCode=0 Dec 03 22:14:45 crc kubenswrapper[4715]: I1203 22:14:45.822893 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" event={"ID":"921209ad-aef3-4407-9bde-3ab12ea68e43","Type":"ContainerDied","Data":"0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e"} Dec 03 22:14:52 crc kubenswrapper[4715]: I1203 22:14:52.889453 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" event={"ID":"921209ad-aef3-4407-9bde-3ab12ea68e43","Type":"ContainerStarted","Data":"f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab"} Dec 03 22:14:52 crc kubenswrapper[4715]: I1203 22:14:52.890179 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:52 crc kubenswrapper[4715]: I1203 22:14:52.892968 4715 generic.go:334] "Generic (PLEG): container finished" podID="28d0a607-f5ba-4e00-9e3d-111f0132026e" containerID="131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b" exitCode=0 Dec 03 22:14:52 crc kubenswrapper[4715]: I1203 22:14:52.893017 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" event={"ID":"28d0a607-f5ba-4e00-9e3d-111f0132026e","Type":"ContainerDied","Data":"131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b"} Dec 03 22:14:52 crc kubenswrapper[4715]: I1203 22:14:52.913889 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" podStartSLOduration=25.505851724 podStartE2EDuration="25.913859274s" podCreationTimestamp="2025-12-03 22:14:27 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.541724733 +0000 UTC m=+1981.284435328" lastFinishedPulling="2025-12-03 22:14:44.949732283 +0000 UTC m=+1981.692442878" observedRunningTime="2025-12-03 22:14:52.912852598 +0000 UTC m=+1989.655563223" watchObservedRunningTime="2025-12-03 22:14:52.913859274 +0000 UTC m=+1989.656569929" Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.902162 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"af93de35-2858-4a66-80a3-b7be3a28ff3f","Type":"ContainerStarted","Data":"7c4f26149049eb6239f77be7a19202daf928cb1bc929ae253f852bc9f7b94029"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.903955 4715 generic.go:334] "Generic (PLEG): container finished" podID="d495583b-de78-4dd5-aa7b-70e15ee3cf0e" containerID="4311f520e85dda4f3776ee4bc7e1413c3997774292b64ddb0619269ce3ad69e9" exitCode=0 Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.904017 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xq7sg" event={"ID":"d495583b-de78-4dd5-aa7b-70e15ee3cf0e","Type":"ContainerDied","Data":"4311f520e85dda4f3776ee4bc7e1413c3997774292b64ddb0619269ce3ad69e9"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.906566 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" event={"ID":"28d0a607-f5ba-4e00-9e3d-111f0132026e","Type":"ContainerStarted","Data":"c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.906737 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.909192 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e514cbad-5355-479e-be39-a625b874551c","Type":"ContainerStarted","Data":"38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.909573 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.913580 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3b024438-6200-45dd-9297-ae2b071dae15","Type":"ContainerStarted","Data":"9eccc4dbafa03128bd2c06b9cffe02c803f07f8204c76a3cfc7ae24832e3a72a"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.915777 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5fzvj" event={"ID":"893e6b97-27e2-4372-8fc6-90562f2781bb","Type":"ContainerStarted","Data":"da7597879fd14c9fa0f24b74d707533f79f764eab6582f9a49a99eeed189385b"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.916436 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-5fzvj" Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.917636 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"695ea890-c971-4f5a-84ca-fbd83fa5d74a","Type":"ContainerStarted","Data":"06d6e69215cea400577ffb461266e4a8d35608dfadc934002e92524d6fcef815"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.920588 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5e2a0d28-27e4-4a86-8bf7-d5007af5162c","Type":"ContainerStarted","Data":"56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.923696 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5ffb2b06-021a-49c0-abfe-6bb5c8acba3d","Type":"ContainerStarted","Data":"7f05b4756a399aef2222ec18ae7f8b0d7730982012e579ba86fb627da94effd4"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.924033 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.927884 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"b487c9d6-b74d-4db6-8ef4-caee02c83106","Type":"ContainerStarted","Data":"3dd2296586f8a640987c8a87828ba1c2ff42d8aef21802261be3765840f1b4db"} Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.973073 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.656780443 podStartE2EDuration="21.973058069s" podCreationTimestamp="2025-12-03 22:14:32 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.376578037 +0000 UTC m=+1981.119288642" lastFinishedPulling="2025-12-03 22:14:51.692855633 +0000 UTC m=+1988.435566268" observedRunningTime="2025-12-03 22:14:53.970606953 +0000 UTC m=+1990.713317548" watchObservedRunningTime="2025-12-03 22:14:53.973058069 +0000 UTC m=+1990.715768664" Dec 03 22:14:53 crc kubenswrapper[4715]: I1203 22:14:53.996702 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" podStartSLOduration=20.065962336 podStartE2EDuration="25.996686871s" podCreationTimestamp="2025-12-03 22:14:28 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.551832173 +0000 UTC m=+1981.294542768" lastFinishedPulling="2025-12-03 22:14:50.482556708 +0000 UTC m=+1987.225267303" observedRunningTime="2025-12-03 22:14:53.989619362 +0000 UTC m=+1990.732329957" watchObservedRunningTime="2025-12-03 22:14:53.996686871 +0000 UTC m=+1990.739397466" Dec 03 22:14:54 crc kubenswrapper[4715]: I1203 22:14:54.012757 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.578192199 podStartE2EDuration="20.012727139s" podCreationTimestamp="2025-12-03 22:14:34 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.539088482 +0000 UTC m=+1981.281799077" lastFinishedPulling="2025-12-03 22:14:52.973623422 +0000 UTC m=+1989.716334017" observedRunningTime="2025-12-03 22:14:54.009550655 +0000 UTC m=+1990.752261260" watchObservedRunningTime="2025-12-03 22:14:54.012727139 +0000 UTC m=+1990.755437734" Dec 03 22:14:54 crc kubenswrapper[4715]: I1203 22:14:54.033153 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-5fzvj" podStartSLOduration=8.86381129 podStartE2EDuration="16.033134916s" podCreationTimestamp="2025-12-03 22:14:38 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.528964182 +0000 UTC m=+1981.271674777" lastFinishedPulling="2025-12-03 22:14:51.698287778 +0000 UTC m=+1988.440998403" observedRunningTime="2025-12-03 22:14:54.028623555 +0000 UTC m=+1990.771334150" watchObservedRunningTime="2025-12-03 22:14:54.033134916 +0000 UTC m=+1990.775845511" Dec 03 22:14:54 crc kubenswrapper[4715]: I1203 22:14:54.937170 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d6205840-7ea5-4060-844a-f6ce9feed200","Type":"ContainerStarted","Data":"f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd"} Dec 03 22:14:54 crc kubenswrapper[4715]: I1203 22:14:54.941222 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xq7sg" event={"ID":"d495583b-de78-4dd5-aa7b-70e15ee3cf0e","Type":"ContainerStarted","Data":"c9084f5663751b05f02071c58f09541b32a18ac1491d10732f5e7d205a2c14de"} Dec 03 22:14:54 crc kubenswrapper[4715]: I1203 22:14:54.941253 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xq7sg" event={"ID":"d495583b-de78-4dd5-aa7b-70e15ee3cf0e","Type":"ContainerStarted","Data":"df372caf7b36ab44f8abca6c647dfc43afaa560fd9409f12e50f809edecb6817"} Dec 03 22:14:54 crc kubenswrapper[4715]: I1203 22:14:54.941861 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:54 crc kubenswrapper[4715]: I1203 22:14:54.941901 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:14:54 crc kubenswrapper[4715]: I1203 22:14:54.990830 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-xq7sg" podStartSLOduration=10.136311347 podStartE2EDuration="16.990798795s" podCreationTimestamp="2025-12-03 22:14:38 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.467918349 +0000 UTC m=+1981.210628954" lastFinishedPulling="2025-12-03 22:14:51.322405767 +0000 UTC m=+1988.065116402" observedRunningTime="2025-12-03 22:14:54.984527406 +0000 UTC m=+1991.727238041" watchObservedRunningTime="2025-12-03 22:14:54.990798795 +0000 UTC m=+1991.733509400" Dec 03 22:14:56 crc kubenswrapper[4715]: I1203 22:14:56.967025 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"af93de35-2858-4a66-80a3-b7be3a28ff3f","Type":"ContainerStarted","Data":"81f0abd1d25cfae8c8840acbe970dee05f1d01e9340db5ee4a23110fd26e40ae"} Dec 03 22:14:56 crc kubenswrapper[4715]: I1203 22:14:56.969974 4715 generic.go:334] "Generic (PLEG): container finished" podID="b487c9d6-b74d-4db6-8ef4-caee02c83106" containerID="3dd2296586f8a640987c8a87828ba1c2ff42d8aef21802261be3765840f1b4db" exitCode=0 Dec 03 22:14:56 crc kubenswrapper[4715]: I1203 22:14:56.970019 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"b487c9d6-b74d-4db6-8ef4-caee02c83106","Type":"ContainerDied","Data":"3dd2296586f8a640987c8a87828ba1c2ff42d8aef21802261be3765840f1b4db"} Dec 03 22:14:56 crc kubenswrapper[4715]: I1203 22:14:56.973384 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"695ea890-c971-4f5a-84ca-fbd83fa5d74a","Type":"ContainerStarted","Data":"76bcebdd838391affadee01b436276bbf8ea40b6869507cfb5ef1944d5623798"} Dec 03 22:14:56 crc kubenswrapper[4715]: I1203 22:14:56.975966 4715 generic.go:334] "Generic (PLEG): container finished" podID="3b024438-6200-45dd-9297-ae2b071dae15" containerID="9eccc4dbafa03128bd2c06b9cffe02c803f07f8204c76a3cfc7ae24832e3a72a" exitCode=0 Dec 03 22:14:56 crc kubenswrapper[4715]: I1203 22:14:56.976618 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3b024438-6200-45dd-9297-ae2b071dae15","Type":"ContainerDied","Data":"9eccc4dbafa03128bd2c06b9cffe02c803f07f8204c76a3cfc7ae24832e3a72a"} Dec 03 22:14:56 crc kubenswrapper[4715]: I1203 22:14:56.995831 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.216351144 podStartE2EDuration="17.994785513s" podCreationTimestamp="2025-12-03 22:14:39 +0000 UTC" firstStartedPulling="2025-12-03 22:14:45.469960565 +0000 UTC m=+1982.212671160" lastFinishedPulling="2025-12-03 22:14:56.248394924 +0000 UTC m=+1992.991105529" observedRunningTime="2025-12-03 22:14:56.994047574 +0000 UTC m=+1993.736758199" watchObservedRunningTime="2025-12-03 22:14:56.994785513 +0000 UTC m=+1993.737496108" Dec 03 22:14:57 crc kubenswrapper[4715]: I1203 22:14:57.025112 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.388415635 podStartE2EDuration="16.025088324s" podCreationTimestamp="2025-12-03 22:14:41 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.611893839 +0000 UTC m=+1981.354604434" lastFinishedPulling="2025-12-03 22:14:56.248566518 +0000 UTC m=+1992.991277123" observedRunningTime="2025-12-03 22:14:57.016852893 +0000 UTC m=+1993.759563558" watchObservedRunningTime="2025-12-03 22:14:57.025088324 +0000 UTC m=+1993.767798949" Dec 03 22:14:57 crc kubenswrapper[4715]: I1203 22:14:57.579651 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:57 crc kubenswrapper[4715]: I1203 22:14:57.579706 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:57 crc kubenswrapper[4715]: I1203 22:14:57.626127 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:57 crc kubenswrapper[4715]: I1203 22:14:57.990991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"b487c9d6-b74d-4db6-8ef4-caee02c83106","Type":"ContainerStarted","Data":"95de3df3aeb48adb6c3e1ec4dedc7fa6dfe79e783de243982fd8cef755dff49b"} Dec 03 22:14:57 crc kubenswrapper[4715]: I1203 22:14:57.996592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3b024438-6200-45dd-9297-ae2b071dae15","Type":"ContainerStarted","Data":"ef7055cd71e4c4404ef7f208fd17c1c1abdda8ff674f36d0d20d73c36cf06d43"} Dec 03 22:14:58 crc kubenswrapper[4715]: I1203 22:14:58.038703 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=20.10152934 podStartE2EDuration="27.038683108s" podCreationTimestamp="2025-12-03 22:14:31 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.382905876 +0000 UTC m=+1981.125616471" lastFinishedPulling="2025-12-03 22:14:51.320059604 +0000 UTC m=+1988.062770239" observedRunningTime="2025-12-03 22:14:58.033326515 +0000 UTC m=+1994.776037150" watchObservedRunningTime="2025-12-03 22:14:58.038683108 +0000 UTC m=+1994.781393703" Dec 03 22:14:58 crc kubenswrapper[4715]: I1203 22:14:58.061551 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 22:14:58 crc kubenswrapper[4715]: I1203 22:14:58.071139 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=21.412521525 podStartE2EDuration="29.071121866s" podCreationTimestamp="2025-12-03 22:14:29 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.244499605 +0000 UTC m=+1980.987210220" lastFinishedPulling="2025-12-03 22:14:51.903099976 +0000 UTC m=+1988.645810561" observedRunningTime="2025-12-03 22:14:58.070601931 +0000 UTC m=+1994.813312526" watchObservedRunningTime="2025-12-03 22:14:58.071121866 +0000 UTC m=+1994.813832491" Dec 03 22:14:58 crc kubenswrapper[4715]: I1203 22:14:58.168811 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:58 crc kubenswrapper[4715]: I1203 22:14:58.643945 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:14:58 crc kubenswrapper[4715]: I1203 22:14:58.708642 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-94s7g"] Dec 03 22:14:58 crc kubenswrapper[4715]: I1203 22:14:58.852102 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:58 crc kubenswrapper[4715]: I1203 22:14:58.895257 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.003822 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" podUID="921209ad-aef3-4407-9bde-3ab12ea68e43" containerName="dnsmasq-dns" containerID="cri-o://f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab" gracePeriod=10 Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.004816 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.054285 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.060186 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.241615 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ddhrm"] Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.243104 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.248918 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.250422 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ddhrm"] Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.287573 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-sgsmm"] Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.288627 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.291277 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-sgsmm"] Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.291655 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.332562 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smct5\" (UniqueName: \"kubernetes.io/projected/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-kube-api-access-smct5\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.332640 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.332694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.332792 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-config\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.414885 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ddhrm"] Dec 03 22:14:59 crc kubenswrapper[4715]: E1203 22:14:59.415425 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-smct5 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" podUID="6c7e7a89-83b7-4c37-9a41-bd1636cda6e6" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.433053 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434230 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/473308fd-dd98-47fd-8a5e-32ab39bd6730-ovs-rundir\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434270 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/473308fd-dd98-47fd-8a5e-32ab39bd6730-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434329 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473308fd-dd98-47fd-8a5e-32ab39bd6730-combined-ca-bundle\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434347 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/473308fd-dd98-47fd-8a5e-32ab39bd6730-ovn-rundir\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434364 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/473308fd-dd98-47fd-8a5e-32ab39bd6730-config\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434383 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434385 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smct5\" (UniqueName: \"kubernetes.io/projected/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-kube-api-access-smct5\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434872 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434909 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434934 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7rwj6\" (UniqueName: \"kubernetes.io/projected/473308fd-dd98-47fd-8a5e-32ab39bd6730-kube-api-access-7rwj6\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.434957 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-config\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.439152 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.439559 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-config\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.442229 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.442433 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.442670 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.442939 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.448434 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-kwxps"] Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.449622 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.457268 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.457616 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-gdn5l" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.461430 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kwxps"] Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.462227 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smct5\" (UniqueName: \"kubernetes.io/projected/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-kube-api-access-smct5\") pod \"dnsmasq-dns-5bf47b49b7-ddhrm\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.468561 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536494 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473308fd-dd98-47fd-8a5e-32ab39bd6730-combined-ca-bundle\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/473308fd-dd98-47fd-8a5e-32ab39bd6730-ovn-rundir\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536639 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/473308fd-dd98-47fd-8a5e-32ab39bd6730-config\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536669 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18d99e58-0606-46f0-91ef-553c75d44ba4-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536727 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7rwj6\" (UniqueName: \"kubernetes.io/projected/473308fd-dd98-47fd-8a5e-32ab39bd6730-kube-api-access-7rwj6\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536755 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18d99e58-0606-46f0-91ef-553c75d44ba4-scripts\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536779 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536821 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536850 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/473308fd-dd98-47fd-8a5e-32ab39bd6730-ovs-rundir\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536870 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-config\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536899 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/473308fd-dd98-47fd-8a5e-32ab39bd6730-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536921 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.536944 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8kbs\" (UniqueName: \"kubernetes.io/projected/18d99e58-0606-46f0-91ef-553c75d44ba4-kube-api-access-n8kbs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.537169 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/473308fd-dd98-47fd-8a5e-32ab39bd6730-ovn-rundir\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.537168 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/473308fd-dd98-47fd-8a5e-32ab39bd6730-ovs-rundir\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.537435 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d99e58-0606-46f0-91ef-553c75d44ba4-config\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.537461 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckmxf\" (UniqueName: \"kubernetes.io/projected/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-kube-api-access-ckmxf\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.537478 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.537523 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-dns-svc\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.537545 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.537769 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/473308fd-dd98-47fd-8a5e-32ab39bd6730-config\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.546266 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473308fd-dd98-47fd-8a5e-32ab39bd6730-combined-ca-bundle\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.548052 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/473308fd-dd98-47fd-8a5e-32ab39bd6730-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.585253 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7rwj6\" (UniqueName: \"kubernetes.io/projected/473308fd-dd98-47fd-8a5e-32ab39bd6730-kube-api-access-7rwj6\") pod \"ovn-controller-metrics-sgsmm\" (UID: \"473308fd-dd98-47fd-8a5e-32ab39bd6730\") " pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.604245 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-sgsmm" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.614674 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639370 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-dns-svc\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639409 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639453 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18d99e58-0606-46f0-91ef-553c75d44ba4-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639518 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18d99e58-0606-46f0-91ef-553c75d44ba4-scripts\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639536 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639584 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-config\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639606 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639622 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8kbs\" (UniqueName: \"kubernetes.io/projected/18d99e58-0606-46f0-91ef-553c75d44ba4-kube-api-access-n8kbs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639646 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d99e58-0606-46f0-91ef-553c75d44ba4-config\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639664 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckmxf\" (UniqueName: \"kubernetes.io/projected/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-kube-api-access-ckmxf\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.639679 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.640164 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-dns-svc\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.641770 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.641969 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-config\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.642115 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.642652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/18d99e58-0606-46f0-91ef-553c75d44ba4-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.642838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18d99e58-0606-46f0-91ef-553c75d44ba4-scripts\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.643438 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18d99e58-0606-46f0-91ef-553c75d44ba4-config\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.648071 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.649937 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.656330 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/18d99e58-0606-46f0-91ef-553c75d44ba4-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.657000 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8kbs\" (UniqueName: \"kubernetes.io/projected/18d99e58-0606-46f0-91ef-553c75d44ba4-kube-api-access-n8kbs\") pod \"ovn-northd-0\" (UID: \"18d99e58-0606-46f0-91ef-553c75d44ba4\") " pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.658849 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckmxf\" (UniqueName: \"kubernetes.io/projected/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-kube-api-access-ckmxf\") pod \"dnsmasq-dns-8554648995-kwxps\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.741246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmdvs\" (UniqueName: \"kubernetes.io/projected/921209ad-aef3-4407-9bde-3ab12ea68e43-kube-api-access-mmdvs\") pod \"921209ad-aef3-4407-9bde-3ab12ea68e43\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.741369 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-dns-svc\") pod \"921209ad-aef3-4407-9bde-3ab12ea68e43\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.741399 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-config\") pod \"921209ad-aef3-4407-9bde-3ab12ea68e43\" (UID: \"921209ad-aef3-4407-9bde-3ab12ea68e43\") " Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.754704 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/921209ad-aef3-4407-9bde-3ab12ea68e43-kube-api-access-mmdvs" (OuterVolumeSpecName: "kube-api-access-mmdvs") pod "921209ad-aef3-4407-9bde-3ab12ea68e43" (UID: "921209ad-aef3-4407-9bde-3ab12ea68e43"). InnerVolumeSpecName "kube-api-access-mmdvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.788288 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-config" (OuterVolumeSpecName: "config") pod "921209ad-aef3-4407-9bde-3ab12ea68e43" (UID: "921209ad-aef3-4407-9bde-3ab12ea68e43"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.789117 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "921209ad-aef3-4407-9bde-3ab12ea68e43" (UID: "921209ad-aef3-4407-9bde-3ab12ea68e43"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.795141 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.806539 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.843779 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.843871 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/921209ad-aef3-4407-9bde-3ab12ea68e43-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:14:59 crc kubenswrapper[4715]: I1203 22:14:59.843888 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmdvs\" (UniqueName: \"kubernetes.io/projected/921209ad-aef3-4407-9bde-3ab12ea68e43-kube-api-access-mmdvs\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.028923 4715 generic.go:334] "Generic (PLEG): container finished" podID="921209ad-aef3-4407-9bde-3ab12ea68e43" containerID="f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab" exitCode=0 Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.028997 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" event={"ID":"921209ad-aef3-4407-9bde-3ab12ea68e43","Type":"ContainerDied","Data":"f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab"} Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.029066 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.029101 4715 scope.go:117] "RemoveContainer" containerID="f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.034323 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-94s7g" event={"ID":"921209ad-aef3-4407-9bde-3ab12ea68e43","Type":"ContainerDied","Data":"b2bf99d7bd896eab645122c28bbd3e83fba5a7d2de8d938c5c949769a06f4664"} Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.034693 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.038478 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-sgsmm"] Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.047575 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.092974 4715 scope.go:117] "RemoveContainer" containerID="0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.093193 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-94s7g"] Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.099202 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-94s7g"] Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.132656 4715 scope.go:117] "RemoveContainer" containerID="f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab" Dec 03 22:15:00 crc kubenswrapper[4715]: E1203 22:15:00.133189 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab\": container with ID starting with f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab not found: ID does not exist" containerID="f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.133238 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab"} err="failed to get container status \"f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab\": rpc error: code = NotFound desc = could not find container \"f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab\": container with ID starting with f8ba66f2b6488a292f18818a778101dfe750aa6d3d1a1cf2a5eb26f9e1ab99ab not found: ID does not exist" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.133265 4715 scope.go:117] "RemoveContainer" containerID="0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e" Dec 03 22:15:00 crc kubenswrapper[4715]: E1203 22:15:00.133647 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e\": container with ID starting with 0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e not found: ID does not exist" containerID="0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.133674 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e"} err="failed to get container status \"0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e\": rpc error: code = NotFound desc = could not find container \"0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e\": container with ID starting with 0c0584574a8ceac1065ceba485446ecd26fc2dba635d7879809f8e9eda4ab29e not found: ID does not exist" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.137406 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2"] Dec 03 22:15:00 crc kubenswrapper[4715]: E1203 22:15:00.137978 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="921209ad-aef3-4407-9bde-3ab12ea68e43" containerName="dnsmasq-dns" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.137998 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="921209ad-aef3-4407-9bde-3ab12ea68e43" containerName="dnsmasq-dns" Dec 03 22:15:00 crc kubenswrapper[4715]: E1203 22:15:00.138016 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="921209ad-aef3-4407-9bde-3ab12ea68e43" containerName="init" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.138025 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="921209ad-aef3-4407-9bde-3ab12ea68e43" containerName="init" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.138221 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="921209ad-aef3-4407-9bde-3ab12ea68e43" containerName="dnsmasq-dns" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.144877 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.146175 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2"] Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.148041 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.148094 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.148406 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-ovsdbserver-nb\") pod \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.148471 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smct5\" (UniqueName: \"kubernetes.io/projected/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-kube-api-access-smct5\") pod \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.148581 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-dns-svc\") pod \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.148683 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-config\") pod \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\" (UID: \"6c7e7a89-83b7-4c37-9a41-bd1636cda6e6\") " Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.149273 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6c7e7a89-83b7-4c37-9a41-bd1636cda6e6" (UID: "6c7e7a89-83b7-4c37-9a41-bd1636cda6e6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.150763 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c7e7a89-83b7-4c37-9a41-bd1636cda6e6" (UID: "6c7e7a89-83b7-4c37-9a41-bd1636cda6e6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.151122 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-config" (OuterVolumeSpecName: "config") pod "6c7e7a89-83b7-4c37-9a41-bd1636cda6e6" (UID: "6c7e7a89-83b7-4c37-9a41-bd1636cda6e6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.153175 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-kube-api-access-smct5" (OuterVolumeSpecName: "kube-api-access-smct5") pod "6c7e7a89-83b7-4c37-9a41-bd1636cda6e6" (UID: "6c7e7a89-83b7-4c37-9a41-bd1636cda6e6"). InnerVolumeSpecName "kube-api-access-smct5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.250785 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6579b54-5da2-431c-beb6-e688bc93560c-config-volume\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.251263 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcn8s\" (UniqueName: \"kubernetes.io/projected/e6579b54-5da2-431c-beb6-e688bc93560c-kube-api-access-xcn8s\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.251293 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6579b54-5da2-431c-beb6-e688bc93560c-secret-volume\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.251468 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.251485 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smct5\" (UniqueName: \"kubernetes.io/projected/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-kube-api-access-smct5\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.251518 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.251533 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.264681 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 22:15:00 crc kubenswrapper[4715]: W1203 22:15:00.269698 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18d99e58_0606_46f0_91ef_553c75d44ba4.slice/crio-d7cd60a026baac79362a4f38d142855ccc813c75b85578cf4f73c8bc548c78aa WatchSource:0}: Error finding container d7cd60a026baac79362a4f38d142855ccc813c75b85578cf4f73c8bc548c78aa: Status 404 returned error can't find the container with id d7cd60a026baac79362a4f38d142855ccc813c75b85578cf4f73c8bc548c78aa Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.312342 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kwxps"] Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.352450 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6579b54-5da2-431c-beb6-e688bc93560c-config-volume\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.352530 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcn8s\" (UniqueName: \"kubernetes.io/projected/e6579b54-5da2-431c-beb6-e688bc93560c-kube-api-access-xcn8s\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.352550 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6579b54-5da2-431c-beb6-e688bc93560c-secret-volume\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.353771 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6579b54-5da2-431c-beb6-e688bc93560c-config-volume\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.357582 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6579b54-5da2-431c-beb6-e688bc93560c-secret-volume\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.376955 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcn8s\" (UniqueName: \"kubernetes.io/projected/e6579b54-5da2-431c-beb6-e688bc93560c-kube-api-access-xcn8s\") pod \"collect-profiles-29413335-5pbl2\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:00 crc kubenswrapper[4715]: I1203 22:15:00.568213 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.039106 4715 generic.go:334] "Generic (PLEG): container finished" podID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerID="1c9e1429ac2ddea0fc0f306cd3247f681e87cc57ec8ba2658b001ea2e16f10a8" exitCode=0 Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.039490 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kwxps" event={"ID":"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d","Type":"ContainerDied","Data":"1c9e1429ac2ddea0fc0f306cd3247f681e87cc57ec8ba2658b001ea2e16f10a8"} Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.039529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kwxps" event={"ID":"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d","Type":"ContainerStarted","Data":"194b1eb8cf4eb52889401462195f0dfc03090067b4df2b0c4e116b8691af8966"} Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.042197 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-sgsmm" event={"ID":"473308fd-dd98-47fd-8a5e-32ab39bd6730","Type":"ContainerStarted","Data":"a6f95039e90535116fcd4586df36494822cff1f959654f21aa99eb1a68bfdbff"} Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.042243 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-sgsmm" event={"ID":"473308fd-dd98-47fd-8a5e-32ab39bd6730","Type":"ContainerStarted","Data":"746bee8d42cc4c936b6a2f17f4d1719f83150872d214424103ccf3251dde93a8"} Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.046581 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18d99e58-0606-46f0-91ef-553c75d44ba4","Type":"ContainerStarted","Data":"d7cd60a026baac79362a4f38d142855ccc813c75b85578cf4f73c8bc548c78aa"} Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.046683 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-ddhrm" Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.093034 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2"] Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.101235 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-sgsmm" podStartSLOduration=2.101211923 podStartE2EDuration="2.101211923s" podCreationTimestamp="2025-12-03 22:14:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:15:01.088401661 +0000 UTC m=+1997.831112296" watchObservedRunningTime="2025-12-03 22:15:01.101211923 +0000 UTC m=+1997.843922518" Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.137917 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ddhrm"] Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.150951 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-ddhrm"] Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.224610 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.224667 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.644129 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c7e7a89-83b7-4c37-9a41-bd1636cda6e6" path="/var/lib/kubelet/pods/6c7e7a89-83b7-4c37-9a41-bd1636cda6e6/volumes" Dec 03 22:15:01 crc kubenswrapper[4715]: I1203 22:15:01.644625 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="921209ad-aef3-4407-9bde-3ab12ea68e43" path="/var/lib/kubelet/pods/921209ad-aef3-4407-9bde-3ab12ea68e43/volumes" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.059319 4715 generic.go:334] "Generic (PLEG): container finished" podID="e6579b54-5da2-431c-beb6-e688bc93560c" containerID="c07ff48571ffc753f77a1e86664859a3317ec1ca39f20791fa1aa3d1cbf323b6" exitCode=0 Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.059433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" event={"ID":"e6579b54-5da2-431c-beb6-e688bc93560c","Type":"ContainerDied","Data":"c07ff48571ffc753f77a1e86664859a3317ec1ca39f20791fa1aa3d1cbf323b6"} Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.059875 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" event={"ID":"e6579b54-5da2-431c-beb6-e688bc93560c","Type":"ContainerStarted","Data":"ec59d9a2e55a37b9c431deb75b9ab70fbe3154bc320680a5178796a67dc49328"} Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.063534 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kwxps" event={"ID":"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d","Type":"ContainerStarted","Data":"5946d14eda69caa155d0bf14897ca4df1cd1cf76e3c6e3e724404a4ce73d5542"} Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.065283 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.067994 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18d99e58-0606-46f0-91ef-553c75d44ba4","Type":"ContainerStarted","Data":"7cf02df380fded07ce86c01990952affc28ad95f1b8cdf11caa0156752086047"} Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.068043 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"18d99e58-0606-46f0-91ef-553c75d44ba4","Type":"ContainerStarted","Data":"f6ba9c25057c27db2c7f6c0b16834172d2223ffd4f7394c49d7d36985cca89d0"} Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.068068 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.108068 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-kwxps" podStartSLOduration=3.108048007 podStartE2EDuration="3.108048007s" podCreationTimestamp="2025-12-03 22:14:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:15:02.10405754 +0000 UTC m=+1998.846768145" watchObservedRunningTime="2025-12-03 22:15:02.108048007 +0000 UTC m=+1998.850758612" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.129355 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.138139 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.946211789 podStartE2EDuration="3.138115541s" podCreationTimestamp="2025-12-03 22:14:59 +0000 UTC" firstStartedPulling="2025-12-03 22:15:00.272738709 +0000 UTC m=+1997.015449314" lastFinishedPulling="2025-12-03 22:15:01.464642481 +0000 UTC m=+1998.207353066" observedRunningTime="2025-12-03 22:15:02.12760964 +0000 UTC m=+1998.870320245" watchObservedRunningTime="2025-12-03 22:15:02.138115541 +0000 UTC m=+1998.880826146" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.255038 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.648932 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2344-account-create-update-2zxn7"] Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.652132 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.654688 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.658885 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2344-account-create-update-2zxn7"] Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.689819 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ftwk\" (UniqueName: \"kubernetes.io/projected/4a676b7b-101e-4fd5-8aae-828dc3c502a7-kube-api-access-5ftwk\") pod \"keystone-2344-account-create-update-2zxn7\" (UID: \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\") " pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.690565 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a676b7b-101e-4fd5-8aae-828dc3c502a7-operator-scripts\") pod \"keystone-2344-account-create-update-2zxn7\" (UID: \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\") " pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.716597 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-gc4h7"] Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.718068 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.731613 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gc4h7"] Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.756133 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.756212 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.791766 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a676b7b-101e-4fd5-8aae-828dc3c502a7-operator-scripts\") pod \"keystone-2344-account-create-update-2zxn7\" (UID: \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\") " pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.791879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ftwk\" (UniqueName: \"kubernetes.io/projected/4a676b7b-101e-4fd5-8aae-828dc3c502a7-kube-api-access-5ftwk\") pod \"keystone-2344-account-create-update-2zxn7\" (UID: \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\") " pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.791932 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpf2q\" (UniqueName: \"kubernetes.io/projected/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-kube-api-access-hpf2q\") pod \"keystone-db-create-gc4h7\" (UID: \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\") " pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.791951 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-operator-scripts\") pod \"keystone-db-create-gc4h7\" (UID: \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\") " pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.793133 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a676b7b-101e-4fd5-8aae-828dc3c502a7-operator-scripts\") pod \"keystone-2344-account-create-update-2zxn7\" (UID: \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\") " pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.810319 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ftwk\" (UniqueName: \"kubernetes.io/projected/4a676b7b-101e-4fd5-8aae-828dc3c502a7-kube-api-access-5ftwk\") pod \"keystone-2344-account-create-update-2zxn7\" (UID: \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\") " pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.893417 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpf2q\" (UniqueName: \"kubernetes.io/projected/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-kube-api-access-hpf2q\") pod \"keystone-db-create-gc4h7\" (UID: \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\") " pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.893728 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-operator-scripts\") pod \"keystone-db-create-gc4h7\" (UID: \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\") " pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.894693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-operator-scripts\") pod \"keystone-db-create-gc4h7\" (UID: \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\") " pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.914092 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpf2q\" (UniqueName: \"kubernetes.io/projected/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-kube-api-access-hpf2q\") pod \"keystone-db-create-gc4h7\" (UID: \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\") " pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.947982 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-9js9r"] Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.949082 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9js9r" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.956386 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-1c8c-account-create-update-8rqw4"] Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.957409 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:02 crc kubenswrapper[4715]: I1203 22:15:02.959773 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:02.978045 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1c8c-account-create-update-8rqw4"] Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.009947 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.010687 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-operator-scripts\") pod \"placement-db-create-9js9r\" (UID: \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\") " pod="openstack/placement-db-create-9js9r" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.010753 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hczg7\" (UniqueName: \"kubernetes.io/projected/bdab221f-c97f-45d1-992a-d00b1e862b1a-kube-api-access-hczg7\") pod \"placement-1c8c-account-create-update-8rqw4\" (UID: \"bdab221f-c97f-45d1-992a-d00b1e862b1a\") " pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.010807 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdab221f-c97f-45d1-992a-d00b1e862b1a-operator-scripts\") pod \"placement-1c8c-account-create-update-8rqw4\" (UID: \"bdab221f-c97f-45d1-992a-d00b1e862b1a\") " pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.010841 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm8sr\" (UniqueName: \"kubernetes.io/projected/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-kube-api-access-gm8sr\") pod \"placement-db-create-9js9r\" (UID: \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\") " pod="openstack/placement-db-create-9js9r" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.027861 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-9js9r"] Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.038066 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.113079 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdab221f-c97f-45d1-992a-d00b1e862b1a-operator-scripts\") pod \"placement-1c8c-account-create-update-8rqw4\" (UID: \"bdab221f-c97f-45d1-992a-d00b1e862b1a\") " pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.115104 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm8sr\" (UniqueName: \"kubernetes.io/projected/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-kube-api-access-gm8sr\") pod \"placement-db-create-9js9r\" (UID: \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\") " pod="openstack/placement-db-create-9js9r" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.115536 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdab221f-c97f-45d1-992a-d00b1e862b1a-operator-scripts\") pod \"placement-1c8c-account-create-update-8rqw4\" (UID: \"bdab221f-c97f-45d1-992a-d00b1e862b1a\") " pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.115619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-operator-scripts\") pod \"placement-db-create-9js9r\" (UID: \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\") " pod="openstack/placement-db-create-9js9r" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.117318 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-operator-scripts\") pod \"placement-db-create-9js9r\" (UID: \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\") " pod="openstack/placement-db-create-9js9r" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.126636 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hczg7\" (UniqueName: \"kubernetes.io/projected/bdab221f-c97f-45d1-992a-d00b1e862b1a-kube-api-access-hczg7\") pod \"placement-1c8c-account-create-update-8rqw4\" (UID: \"bdab221f-c97f-45d1-992a-d00b1e862b1a\") " pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.138550 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm8sr\" (UniqueName: \"kubernetes.io/projected/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-kube-api-access-gm8sr\") pod \"placement-db-create-9js9r\" (UID: \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\") " pod="openstack/placement-db-create-9js9r" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.138979 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hczg7\" (UniqueName: \"kubernetes.io/projected/bdab221f-c97f-45d1-992a-d00b1e862b1a-kube-api-access-hczg7\") pod \"placement-1c8c-account-create-update-8rqw4\" (UID: \"bdab221f-c97f-45d1-992a-d00b1e862b1a\") " pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.387949 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9js9r" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.403143 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.464812 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.551085 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-gc4h7"] Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.574517 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2344-account-create-update-2zxn7"] Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.640620 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6579b54-5da2-431c-beb6-e688bc93560c-secret-volume\") pod \"e6579b54-5da2-431c-beb6-e688bc93560c\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.640986 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcn8s\" (UniqueName: \"kubernetes.io/projected/e6579b54-5da2-431c-beb6-e688bc93560c-kube-api-access-xcn8s\") pod \"e6579b54-5da2-431c-beb6-e688bc93560c\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.641013 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6579b54-5da2-431c-beb6-e688bc93560c-config-volume\") pod \"e6579b54-5da2-431c-beb6-e688bc93560c\" (UID: \"e6579b54-5da2-431c-beb6-e688bc93560c\") " Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.642064 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6579b54-5da2-431c-beb6-e688bc93560c-config-volume" (OuterVolumeSpecName: "config-volume") pod "e6579b54-5da2-431c-beb6-e688bc93560c" (UID: "e6579b54-5da2-431c-beb6-e688bc93560c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.652281 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6579b54-5da2-431c-beb6-e688bc93560c-kube-api-access-xcn8s" (OuterVolumeSpecName: "kube-api-access-xcn8s") pod "e6579b54-5da2-431c-beb6-e688bc93560c" (UID: "e6579b54-5da2-431c-beb6-e688bc93560c"). InnerVolumeSpecName "kube-api-access-xcn8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.660543 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6579b54-5da2-431c-beb6-e688bc93560c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e6579b54-5da2-431c-beb6-e688bc93560c" (UID: "e6579b54-5da2-431c-beb6-e688bc93560c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.742566 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcn8s\" (UniqueName: \"kubernetes.io/projected/e6579b54-5da2-431c-beb6-e688bc93560c-kube-api-access-xcn8s\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.742595 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e6579b54-5da2-431c-beb6-e688bc93560c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.742604 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e6579b54-5da2-431c-beb6-e688bc93560c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.852749 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-9js9r"] Dec 03 22:15:03 crc kubenswrapper[4715]: W1203 22:15:03.855562 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ab729b6_f4d1_4f48_bb28_43d4c4c81e58.slice/crio-f15037d078287e804279694f38fd082a0abf9fc8ea0cf42ce4d7fc69e3e93d7b WatchSource:0}: Error finding container f15037d078287e804279694f38fd082a0abf9fc8ea0cf42ce4d7fc69e3e93d7b: Status 404 returned error can't find the container with id f15037d078287e804279694f38fd082a0abf9fc8ea0cf42ce4d7fc69e3e93d7b Dec 03 22:15:03 crc kubenswrapper[4715]: I1203 22:15:03.948978 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1c8c-account-create-update-8rqw4"] Dec 03 22:15:04 crc kubenswrapper[4715]: W1203 22:15:04.042956 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdab221f_c97f_45d1_992a_d00b1e862b1a.slice/crio-9f24238e02a85b56318f6a697ee1a97dbb05191c61597e84091c357565330df0 WatchSource:0}: Error finding container 9f24238e02a85b56318f6a697ee1a97dbb05191c61597e84091c357565330df0: Status 404 returned error can't find the container with id 9f24238e02a85b56318f6a697ee1a97dbb05191c61597e84091c357565330df0 Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.091780 4715 generic.go:334] "Generic (PLEG): container finished" podID="4a676b7b-101e-4fd5-8aae-828dc3c502a7" containerID="98b64c6fa621fe81b53455bf9c664b08983758898a8640904a945e9e338a7868" exitCode=0 Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.091891 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2344-account-create-update-2zxn7" event={"ID":"4a676b7b-101e-4fd5-8aae-828dc3c502a7","Type":"ContainerDied","Data":"98b64c6fa621fe81b53455bf9c664b08983758898a8640904a945e9e338a7868"} Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.091921 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2344-account-create-update-2zxn7" event={"ID":"4a676b7b-101e-4fd5-8aae-828dc3c502a7","Type":"ContainerStarted","Data":"3c6b620e6c135745269653a4143fe4f42214ab58764e1a4f7587eed7a503c3fc"} Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.094097 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9js9r" event={"ID":"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58","Type":"ContainerStarted","Data":"f15037d078287e804279694f38fd082a0abf9fc8ea0cf42ce4d7fc69e3e93d7b"} Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.097138 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.097163 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2" event={"ID":"e6579b54-5da2-431c-beb6-e688bc93560c","Type":"ContainerDied","Data":"ec59d9a2e55a37b9c431deb75b9ab70fbe3154bc320680a5178796a67dc49328"} Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.097221 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec59d9a2e55a37b9c431deb75b9ab70fbe3154bc320680a5178796a67dc49328" Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.100309 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1c8c-account-create-update-8rqw4" event={"ID":"bdab221f-c97f-45d1-992a-d00b1e862b1a","Type":"ContainerStarted","Data":"9f24238e02a85b56318f6a697ee1a97dbb05191c61597e84091c357565330df0"} Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.104540 4715 generic.go:334] "Generic (PLEG): container finished" podID="efe6fd2a-8278-4fce-b29f-2f2237c5abf6" containerID="2c508d00b0d0d5fbc4e608b3df2d31feb8cc94c11cc4ce0a099ed3beea3aba97" exitCode=0 Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.104585 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gc4h7" event={"ID":"efe6fd2a-8278-4fce-b29f-2f2237c5abf6","Type":"ContainerDied","Data":"2c508d00b0d0d5fbc4e608b3df2d31feb8cc94c11cc4ce0a099ed3beea3aba97"} Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.104612 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gc4h7" event={"ID":"efe6fd2a-8278-4fce-b29f-2f2237c5abf6","Type":"ContainerStarted","Data":"f0fc7484385269924ea8df615c868656ee01067ebaaaffdd488f6082f7c4d349"} Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.540100 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc"] Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.550113 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413290-nwfsc"] Dec 03 22:15:04 crc kubenswrapper[4715]: I1203 22:15:04.985356 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.056474 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.061654 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kwxps"] Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.061883 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-kwxps" podUID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerName="dnsmasq-dns" containerID="cri-o://5946d14eda69caa155d0bf14897ca4df1cd1cf76e3c6e3e724404a4ce73d5542" gracePeriod=10 Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.119207 4715 generic.go:334] "Generic (PLEG): container finished" podID="bdab221f-c97f-45d1-992a-d00b1e862b1a" containerID="b35a24b8a7c11936af70b902121801d8d56caadfb1006683f835e78a475872fe" exitCode=0 Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.119279 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1c8c-account-create-update-8rqw4" event={"ID":"bdab221f-c97f-45d1-992a-d00b1e862b1a","Type":"ContainerDied","Data":"b35a24b8a7c11936af70b902121801d8d56caadfb1006683f835e78a475872fe"} Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.128170 4715 generic.go:334] "Generic (PLEG): container finished" podID="4ab729b6-f4d1-4f48-bb28-43d4c4c81e58" containerID="810ee17edc7caf608374db8da1fc528a65b67b80c0a831e8b543e7649025f8e5" exitCode=0 Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.128444 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9js9r" event={"ID":"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58","Type":"ContainerDied","Data":"810ee17edc7caf608374db8da1fc528a65b67b80c0a831e8b543e7649025f8e5"} Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.130842 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bqnl2"] Dec 03 22:15:05 crc kubenswrapper[4715]: E1203 22:15:05.131230 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6579b54-5da2-431c-beb6-e688bc93560c" containerName="collect-profiles" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.131253 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6579b54-5da2-431c-beb6-e688bc93560c" containerName="collect-profiles" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.131488 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6579b54-5da2-431c-beb6-e688bc93560c" containerName="collect-profiles" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.133047 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.160074 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bqnl2"] Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.163772 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.163889 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.205656 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.291686 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-config\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.291735 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.291811 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8zlt\" (UniqueName: \"kubernetes.io/projected/793fb19c-ce52-40e2-b874-d0c42cf1ad50-kube-api-access-n8zlt\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.291869 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.291934 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.393483 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8zlt\" (UniqueName: \"kubernetes.io/projected/793fb19c-ce52-40e2-b874-d0c42cf1ad50-kube-api-access-n8zlt\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.393585 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.393645 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.393794 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-config\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.393819 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.396815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.397826 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.398084 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.398101 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-config\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.431892 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8zlt\" (UniqueName: \"kubernetes.io/projected/793fb19c-ce52-40e2-b874-d0c42cf1ad50-kube-api-access-n8zlt\") pod \"dnsmasq-dns-b8fbc5445-bqnl2\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.461050 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.631124 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.652419 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d88e13c-7334-49a9-822b-9ac54d6bf72f" path="/var/lib/kubelet/pods/7d88e13c-7334-49a9-822b-9ac54d6bf72f/volumes" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.757246 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.808330 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpf2q\" (UniqueName: \"kubernetes.io/projected/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-kube-api-access-hpf2q\") pod \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\" (UID: \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\") " Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.808603 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-operator-scripts\") pod \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\" (UID: \"efe6fd2a-8278-4fce-b29f-2f2237c5abf6\") " Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.809275 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "efe6fd2a-8278-4fce-b29f-2f2237c5abf6" (UID: "efe6fd2a-8278-4fce-b29f-2f2237c5abf6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.810452 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.811749 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-kube-api-access-hpf2q" (OuterVolumeSpecName: "kube-api-access-hpf2q") pod "efe6fd2a-8278-4fce-b29f-2f2237c5abf6" (UID: "efe6fd2a-8278-4fce-b29f-2f2237c5abf6"). InnerVolumeSpecName "kube-api-access-hpf2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.911384 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ftwk\" (UniqueName: \"kubernetes.io/projected/4a676b7b-101e-4fd5-8aae-828dc3c502a7-kube-api-access-5ftwk\") pod \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\" (UID: \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\") " Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.911472 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a676b7b-101e-4fd5-8aae-828dc3c502a7-operator-scripts\") pod \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\" (UID: \"4a676b7b-101e-4fd5-8aae-828dc3c502a7\") " Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.911804 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpf2q\" (UniqueName: \"kubernetes.io/projected/efe6fd2a-8278-4fce-b29f-2f2237c5abf6-kube-api-access-hpf2q\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.911957 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a676b7b-101e-4fd5-8aae-828dc3c502a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4a676b7b-101e-4fd5-8aae-828dc3c502a7" (UID: "4a676b7b-101e-4fd5-8aae-828dc3c502a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:05 crc kubenswrapper[4715]: I1203 22:15:05.927202 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a676b7b-101e-4fd5-8aae-828dc3c502a7-kube-api-access-5ftwk" (OuterVolumeSpecName: "kube-api-access-5ftwk") pod "4a676b7b-101e-4fd5-8aae-828dc3c502a7" (UID: "4a676b7b-101e-4fd5-8aae-828dc3c502a7"). InnerVolumeSpecName "kube-api-access-5ftwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.013073 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ftwk\" (UniqueName: \"kubernetes.io/projected/4a676b7b-101e-4fd5-8aae-828dc3c502a7-kube-api-access-5ftwk\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.013111 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a676b7b-101e-4fd5-8aae-828dc3c502a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.014432 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bqnl2"] Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.137749 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2344-account-create-update-2zxn7" event={"ID":"4a676b7b-101e-4fd5-8aae-828dc3c502a7","Type":"ContainerDied","Data":"3c6b620e6c135745269653a4143fe4f42214ab58764e1a4f7587eed7a503c3fc"} Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.138136 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c6b620e6c135745269653a4143fe4f42214ab58764e1a4f7587eed7a503c3fc" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.137989 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2344-account-create-update-2zxn7" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.139537 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" event={"ID":"793fb19c-ce52-40e2-b874-d0c42cf1ad50","Type":"ContainerStarted","Data":"076ffb1543443d2cb189099533b094e6217f904304eed93f0afec356105ba44c"} Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.142650 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-gc4h7" event={"ID":"efe6fd2a-8278-4fce-b29f-2f2237c5abf6","Type":"ContainerDied","Data":"f0fc7484385269924ea8df615c868656ee01067ebaaaffdd488f6082f7c4d349"} Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.142722 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0fc7484385269924ea8df615c868656ee01067ebaaaffdd488f6082f7c4d349" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.142723 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-gc4h7" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.144619 4715 generic.go:334] "Generic (PLEG): container finished" podID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerID="5946d14eda69caa155d0bf14897ca4df1cd1cf76e3c6e3e724404a4ce73d5542" exitCode=0 Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.144728 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kwxps" event={"ID":"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d","Type":"ContainerDied","Data":"5946d14eda69caa155d0bf14897ca4df1cd1cf76e3c6e3e724404a4ce73d5542"} Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.317135 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 22:15:06 crc kubenswrapper[4715]: E1203 22:15:06.317512 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe6fd2a-8278-4fce-b29f-2f2237c5abf6" containerName="mariadb-database-create" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.321340 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe6fd2a-8278-4fce-b29f-2f2237c5abf6" containerName="mariadb-database-create" Dec 03 22:15:06 crc kubenswrapper[4715]: E1203 22:15:06.321658 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a676b7b-101e-4fd5-8aae-828dc3c502a7" containerName="mariadb-account-create-update" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.321666 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a676b7b-101e-4fd5-8aae-828dc3c502a7" containerName="mariadb-account-create-update" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.321846 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a676b7b-101e-4fd5-8aae-828dc3c502a7" containerName="mariadb-account-create-update" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.321867 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe6fd2a-8278-4fce-b29f-2f2237c5abf6" containerName="mariadb-database-create" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.327451 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.330364 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-9tgpl" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.330653 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.330792 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.335007 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.345638 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.520858 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.520924 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d58c9eac-e46a-4a54-b1eb-27246d9c064e-cache\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.521042 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqzz2\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-kube-api-access-zqzz2\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.521076 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d58c9eac-e46a-4a54-b1eb-27246d9c064e-lock\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.521111 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.572727 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9js9r" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.578237 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.622931 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d58c9eac-e46a-4a54-b1eb-27246d9c064e-cache\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.623028 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqzz2\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-kube-api-access-zqzz2\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.623070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d58c9eac-e46a-4a54-b1eb-27246d9c064e-lock\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.623094 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.623155 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: E1203 22:15:06.623312 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 22:15:06 crc kubenswrapper[4715]: E1203 22:15:06.623328 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 22:15:06 crc kubenswrapper[4715]: E1203 22:15:06.623381 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift podName:d58c9eac-e46a-4a54-b1eb-27246d9c064e nodeName:}" failed. No retries permitted until 2025-12-03 22:15:07.123365882 +0000 UTC m=+2003.866076477 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift") pod "swift-storage-0" (UID: "d58c9eac-e46a-4a54-b1eb-27246d9c064e") : configmap "swift-ring-files" not found Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.623496 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/d58c9eac-e46a-4a54-b1eb-27246d9c064e-cache\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.623667 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/d58c9eac-e46a-4a54-b1eb-27246d9c064e-lock\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.623858 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.646168 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqzz2\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-kube-api-access-zqzz2\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.647306 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.733257 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-operator-scripts\") pod \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\" (UID: \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\") " Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.744963 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm8sr\" (UniqueName: \"kubernetes.io/projected/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-kube-api-access-gm8sr\") pod \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\" (UID: \"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58\") " Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.745036 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hczg7\" (UniqueName: \"kubernetes.io/projected/bdab221f-c97f-45d1-992a-d00b1e862b1a-kube-api-access-hczg7\") pod \"bdab221f-c97f-45d1-992a-d00b1e862b1a\" (UID: \"bdab221f-c97f-45d1-992a-d00b1e862b1a\") " Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.745094 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdab221f-c97f-45d1-992a-d00b1e862b1a-operator-scripts\") pod \"bdab221f-c97f-45d1-992a-d00b1e862b1a\" (UID: \"bdab221f-c97f-45d1-992a-d00b1e862b1a\") " Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.746936 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ab729b6-f4d1-4f48-bb28-43d4c4c81e58" (UID: "4ab729b6-f4d1-4f48-bb28-43d4c4c81e58"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.747191 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdab221f-c97f-45d1-992a-d00b1e862b1a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bdab221f-c97f-45d1-992a-d00b1e862b1a" (UID: "bdab221f-c97f-45d1-992a-d00b1e862b1a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.807320 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-kube-api-access-gm8sr" (OuterVolumeSpecName: "kube-api-access-gm8sr") pod "4ab729b6-f4d1-4f48-bb28-43d4c4c81e58" (UID: "4ab729b6-f4d1-4f48-bb28-43d4c4c81e58"). InnerVolumeSpecName "kube-api-access-gm8sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.807719 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdab221f-c97f-45d1-992a-d00b1e862b1a-kube-api-access-hczg7" (OuterVolumeSpecName: "kube-api-access-hczg7") pod "bdab221f-c97f-45d1-992a-d00b1e862b1a" (UID: "bdab221f-c97f-45d1-992a-d00b1e862b1a"). InnerVolumeSpecName "kube-api-access-hczg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.846750 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdab221f-c97f-45d1-992a-d00b1e862b1a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.846787 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.846797 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm8sr\" (UniqueName: \"kubernetes.io/projected/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58-kube-api-access-gm8sr\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.846807 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hczg7\" (UniqueName: \"kubernetes.io/projected/bdab221f-c97f-45d1-992a-d00b1e862b1a-kube-api-access-hczg7\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.910624 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-7f2gr"] Dec 03 22:15:06 crc kubenswrapper[4715]: E1203 22:15:06.910968 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdab221f-c97f-45d1-992a-d00b1e862b1a" containerName="mariadb-account-create-update" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.910986 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdab221f-c97f-45d1-992a-d00b1e862b1a" containerName="mariadb-account-create-update" Dec 03 22:15:06 crc kubenswrapper[4715]: E1203 22:15:06.911000 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab729b6-f4d1-4f48-bb28-43d4c4c81e58" containerName="mariadb-database-create" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.911006 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab729b6-f4d1-4f48-bb28-43d4c4c81e58" containerName="mariadb-database-create" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.911164 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdab221f-c97f-45d1-992a-d00b1e862b1a" containerName="mariadb-account-create-update" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.911188 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab729b6-f4d1-4f48-bb28-43d4c4c81e58" containerName="mariadb-database-create" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.911774 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.914791 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.916050 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.916253 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 22:15:06 crc kubenswrapper[4715]: I1203 22:15:06.938921 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7f2gr"] Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.051666 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-swiftconf\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.051713 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-ring-data-devices\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.051772 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-dispersionconf\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.051788 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-combined-ca-bundle\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.051810 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-258tq\" (UniqueName: \"kubernetes.io/projected/18eb44f1-da35-4a10-b826-03c94f6b525e-kube-api-access-258tq\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.051839 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/18eb44f1-da35-4a10-b826-03c94f6b525e-etc-swift\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.051863 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-scripts\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.153179 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-9js9r" event={"ID":"4ab729b6-f4d1-4f48-bb28-43d4c4c81e58","Type":"ContainerDied","Data":"f15037d078287e804279694f38fd082a0abf9fc8ea0cf42ce4d7fc69e3e93d7b"} Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.153235 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f15037d078287e804279694f38fd082a0abf9fc8ea0cf42ce4d7fc69e3e93d7b" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.153235 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.153280 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-dispersionconf\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.153189 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-9js9r" Dec 03 22:15:07 crc kubenswrapper[4715]: E1203 22:15:07.153429 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 22:15:07 crc kubenswrapper[4715]: E1203 22:15:07.153446 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 22:15:07 crc kubenswrapper[4715]: E1203 22:15:07.153498 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift podName:d58c9eac-e46a-4a54-b1eb-27246d9c064e nodeName:}" failed. No retries permitted until 2025-12-03 22:15:08.153477018 +0000 UTC m=+2004.896187623 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift") pod "swift-storage-0" (UID: "d58c9eac-e46a-4a54-b1eb-27246d9c064e") : configmap "swift-ring-files" not found Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.153873 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-combined-ca-bundle\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.154139 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-258tq\" (UniqueName: \"kubernetes.io/projected/18eb44f1-da35-4a10-b826-03c94f6b525e-kube-api-access-258tq\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.154222 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/18eb44f1-da35-4a10-b826-03c94f6b525e-etc-swift\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.154278 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-scripts\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.154608 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-swiftconf\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.154648 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-ring-data-devices\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.155481 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-ring-data-devices\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.157033 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/18eb44f1-da35-4a10-b826-03c94f6b525e-etc-swift\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.157772 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-dispersionconf\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.157837 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1c8c-account-create-update-8rqw4" event={"ID":"bdab221f-c97f-45d1-992a-d00b1e862b1a","Type":"ContainerDied","Data":"9f24238e02a85b56318f6a697ee1a97dbb05191c61597e84091c357565330df0"} Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.157853 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1c8c-account-create-update-8rqw4" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.157862 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f24238e02a85b56318f6a697ee1a97dbb05191c61597e84091c357565330df0" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.159838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-swiftconf\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.160347 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-combined-ca-bundle\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.172958 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-scripts\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.176557 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-258tq\" (UniqueName: \"kubernetes.io/projected/18eb44f1-da35-4a10-b826-03c94f6b525e-kube-api-access-258tq\") pod \"swift-ring-rebalance-7f2gr\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.235078 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:07 crc kubenswrapper[4715]: I1203 22:15:07.710549 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-7f2gr"] Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.166638 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7f2gr" event={"ID":"18eb44f1-da35-4a10-b826-03c94f6b525e","Type":"ContainerStarted","Data":"cbcaf88329e50ad9332759d53a9236a069439aaa7dfdab1af3c19bd57252aaaa"} Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.170824 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:08 crc kubenswrapper[4715]: E1203 22:15:08.171178 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 22:15:08 crc kubenswrapper[4715]: E1203 22:15:08.171231 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 22:15:08 crc kubenswrapper[4715]: E1203 22:15:08.171325 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift podName:d58c9eac-e46a-4a54-b1eb-27246d9c064e nodeName:}" failed. No retries permitted until 2025-12-03 22:15:10.171294575 +0000 UTC m=+2006.914005170 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift") pod "swift-storage-0" (UID: "d58c9eac-e46a-4a54-b1eb-27246d9c064e") : configmap "swift-ring-files" not found Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.321528 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-9l4gg"] Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.323026 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.336571 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9l4gg"] Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.437892 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-eeb1-account-create-update-hg4zn"] Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.439203 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.441332 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.446738 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-eeb1-account-create-update-hg4zn"] Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.487193 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thswv\" (UniqueName: \"kubernetes.io/projected/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-kube-api-access-thswv\") pod \"glance-db-create-9l4gg\" (UID: \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\") " pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.487476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-operator-scripts\") pod \"glance-db-create-9l4gg\" (UID: \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\") " pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.589448 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-operator-scripts\") pod \"glance-db-create-9l4gg\" (UID: \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\") " pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.589566 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thswv\" (UniqueName: \"kubernetes.io/projected/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-kube-api-access-thswv\") pod \"glance-db-create-9l4gg\" (UID: \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\") " pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.589606 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/5872f688-95bb-4443-ba2d-f55ca9b05e97-kube-api-access-9ltzp\") pod \"glance-eeb1-account-create-update-hg4zn\" (UID: \"5872f688-95bb-4443-ba2d-f55ca9b05e97\") " pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.589721 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5872f688-95bb-4443-ba2d-f55ca9b05e97-operator-scripts\") pod \"glance-eeb1-account-create-update-hg4zn\" (UID: \"5872f688-95bb-4443-ba2d-f55ca9b05e97\") " pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.590485 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-operator-scripts\") pod \"glance-db-create-9l4gg\" (UID: \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\") " pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.613370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thswv\" (UniqueName: \"kubernetes.io/projected/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-kube-api-access-thswv\") pod \"glance-db-create-9l4gg\" (UID: \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\") " pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.692718 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/5872f688-95bb-4443-ba2d-f55ca9b05e97-kube-api-access-9ltzp\") pod \"glance-eeb1-account-create-update-hg4zn\" (UID: \"5872f688-95bb-4443-ba2d-f55ca9b05e97\") " pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.692853 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5872f688-95bb-4443-ba2d-f55ca9b05e97-operator-scripts\") pod \"glance-eeb1-account-create-update-hg4zn\" (UID: \"5872f688-95bb-4443-ba2d-f55ca9b05e97\") " pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.693912 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.693935 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5872f688-95bb-4443-ba2d-f55ca9b05e97-operator-scripts\") pod \"glance-eeb1-account-create-update-hg4zn\" (UID: \"5872f688-95bb-4443-ba2d-f55ca9b05e97\") " pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.719342 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/5872f688-95bb-4443-ba2d-f55ca9b05e97-kube-api-access-9ltzp\") pod \"glance-eeb1-account-create-update-hg4zn\" (UID: \"5872f688-95bb-4443-ba2d-f55ca9b05e97\") " pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:08 crc kubenswrapper[4715]: I1203 22:15:08.752193 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:09 crc kubenswrapper[4715]: W1203 22:15:09.217081 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5046d57b_4ce3_4f2c_8cbc_0dee5590b458.slice/crio-6f83ca5c19293c9af6936dd11315dbad3862b25157a23c26df1fa33c4f566212 WatchSource:0}: Error finding container 6f83ca5c19293c9af6936dd11315dbad3862b25157a23c26df1fa33c4f566212: Status 404 returned error can't find the container with id 6f83ca5c19293c9af6936dd11315dbad3862b25157a23c26df1fa33c4f566212 Dec 03 22:15:09 crc kubenswrapper[4715]: I1203 22:15:09.218338 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9l4gg"] Dec 03 22:15:09 crc kubenswrapper[4715]: I1203 22:15:09.291750 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-eeb1-account-create-update-hg4zn"] Dec 03 22:15:09 crc kubenswrapper[4715]: I1203 22:15:09.809231 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-kwxps" podUID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 03 22:15:10 crc kubenswrapper[4715]: I1203 22:15:10.183312 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9l4gg" event={"ID":"5046d57b-4ce3-4f2c-8cbc-0dee5590b458","Type":"ContainerStarted","Data":"6f83ca5c19293c9af6936dd11315dbad3862b25157a23c26df1fa33c4f566212"} Dec 03 22:15:10 crc kubenswrapper[4715]: I1203 22:15:10.184497 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eeb1-account-create-update-hg4zn" event={"ID":"5872f688-95bb-4443-ba2d-f55ca9b05e97","Type":"ContainerStarted","Data":"f865679f8bc3c5925e6771a6dc52e3a484e5bc23f0b92727148227cad163cf17"} Dec 03 22:15:10 crc kubenswrapper[4715]: I1203 22:15:10.225026 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:10 crc kubenswrapper[4715]: E1203 22:15:10.225221 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 22:15:10 crc kubenswrapper[4715]: E1203 22:15:10.225245 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 22:15:10 crc kubenswrapper[4715]: E1203 22:15:10.225311 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift podName:d58c9eac-e46a-4a54-b1eb-27246d9c064e nodeName:}" failed. No retries permitted until 2025-12-03 22:15:14.225289441 +0000 UTC m=+2010.968000036 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift") pod "swift-storage-0" (UID: "d58c9eac-e46a-4a54-b1eb-27246d9c064e") : configmap "swift-ring-files" not found Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.211626 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eeb1-account-create-update-hg4zn" event={"ID":"5872f688-95bb-4443-ba2d-f55ca9b05e97","Type":"ContainerStarted","Data":"4c59cd8db222e8055fd4a086fb0bf9cd2d0bde2960ada2dddcd98d8b19da62d7"} Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.215125 4715 generic.go:334] "Generic (PLEG): container finished" podID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerID="e9c9503666ff888202257ffb5e64214f9ec6908e5c1680cb6d1ea99842a92cf2" exitCode=0 Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.215163 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" event={"ID":"793fb19c-ce52-40e2-b874-d0c42cf1ad50","Type":"ContainerDied","Data":"e9c9503666ff888202257ffb5e64214f9ec6908e5c1680cb6d1ea99842a92cf2"} Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.217716 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9l4gg" event={"ID":"5046d57b-4ce3-4f2c-8cbc-0dee5590b458","Type":"ContainerStarted","Data":"0b2a39c3654d7bea09c4da32a22208bf980028daf9b82fa49d48e4d8cf048b2d"} Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.239849 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-eeb1-account-create-update-hg4zn" podStartSLOduration=5.239820903 podStartE2EDuration="5.239820903s" podCreationTimestamp="2025-12-03 22:15:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:15:13.225995004 +0000 UTC m=+2009.968705619" watchObservedRunningTime="2025-12-03 22:15:13.239820903 +0000 UTC m=+2009.982531508" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.273344 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-9l4gg" podStartSLOduration=5.273324499 podStartE2EDuration="5.273324499s" podCreationTimestamp="2025-12-03 22:15:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:15:13.271974543 +0000 UTC m=+2010.014685148" watchObservedRunningTime="2025-12-03 22:15:13.273324499 +0000 UTC m=+2010.016035104" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.497282 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.655031 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-sb\") pod \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.655103 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-dns-svc\") pod \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.655633 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-config\") pod \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.655700 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-nb\") pod \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.655745 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckmxf\" (UniqueName: \"kubernetes.io/projected/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-kube-api-access-ckmxf\") pod \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\" (UID: \"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d\") " Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.660827 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-kube-api-access-ckmxf" (OuterVolumeSpecName: "kube-api-access-ckmxf") pod "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" (UID: "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d"). InnerVolumeSpecName "kube-api-access-ckmxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.711624 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" (UID: "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.714353 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" (UID: "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.732918 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" (UID: "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.758307 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-config" (OuterVolumeSpecName: "config") pod "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" (UID: "3e99e0a7-7ad8-4700-a5d1-000b78a2d03d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.758982 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.759012 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.759025 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.759041 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckmxf\" (UniqueName: \"kubernetes.io/projected/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-kube-api-access-ckmxf\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:13 crc kubenswrapper[4715]: I1203 22:15:13.759053 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.228122 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" event={"ID":"793fb19c-ce52-40e2-b874-d0c42cf1ad50","Type":"ContainerStarted","Data":"52e2c8febb319080e0af0f7ffc895517cb374371191d68ea7055ba4ea833dfb5"} Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.228587 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.231290 4715 generic.go:334] "Generic (PLEG): container finished" podID="5046d57b-4ce3-4f2c-8cbc-0dee5590b458" containerID="0b2a39c3654d7bea09c4da32a22208bf980028daf9b82fa49d48e4d8cf048b2d" exitCode=0 Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.231370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9l4gg" event={"ID":"5046d57b-4ce3-4f2c-8cbc-0dee5590b458","Type":"ContainerDied","Data":"0b2a39c3654d7bea09c4da32a22208bf980028daf9b82fa49d48e4d8cf048b2d"} Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.241464 4715 generic.go:334] "Generic (PLEG): container finished" podID="5872f688-95bb-4443-ba2d-f55ca9b05e97" containerID="4c59cd8db222e8055fd4a086fb0bf9cd2d0bde2960ada2dddcd98d8b19da62d7" exitCode=0 Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.241634 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eeb1-account-create-update-hg4zn" event={"ID":"5872f688-95bb-4443-ba2d-f55ca9b05e97","Type":"ContainerDied","Data":"4c59cd8db222e8055fd4a086fb0bf9cd2d0bde2960ada2dddcd98d8b19da62d7"} Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.248208 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" podStartSLOduration=9.248182239 podStartE2EDuration="9.248182239s" podCreationTimestamp="2025-12-03 22:15:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:15:14.247057718 +0000 UTC m=+2010.989768353" watchObservedRunningTime="2025-12-03 22:15:14.248182239 +0000 UTC m=+2010.990892874" Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.253166 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-kwxps" event={"ID":"3e99e0a7-7ad8-4700-a5d1-000b78a2d03d","Type":"ContainerDied","Data":"194b1eb8cf4eb52889401462195f0dfc03090067b4df2b0c4e116b8691af8966"} Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.253218 4715 scope.go:117] "RemoveContainer" containerID="5946d14eda69caa155d0bf14897ca4df1cd1cf76e3c6e3e724404a4ce73d5542" Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.253347 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-kwxps" Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.268710 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:14 crc kubenswrapper[4715]: E1203 22:15:14.269385 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 22:15:14 crc kubenswrapper[4715]: E1203 22:15:14.269407 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 22:15:14 crc kubenswrapper[4715]: E1203 22:15:14.269443 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift podName:d58c9eac-e46a-4a54-b1eb-27246d9c064e nodeName:}" failed. No retries permitted until 2025-12-03 22:15:22.269429787 +0000 UTC m=+2019.012140382 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift") pod "swift-storage-0" (UID: "d58c9eac-e46a-4a54-b1eb-27246d9c064e") : configmap "swift-ring-files" not found Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.347537 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kwxps"] Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.361964 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-kwxps"] Dec 03 22:15:14 crc kubenswrapper[4715]: I1203 22:15:14.875557 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 22:15:15 crc kubenswrapper[4715]: I1203 22:15:15.644466 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" path="/var/lib/kubelet/pods/3e99e0a7-7ad8-4700-a5d1-000b78a2d03d/volumes" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.170401 4715 scope.go:117] "RemoveContainer" containerID="1c9e1429ac2ddea0fc0f306cd3247f681e87cc57ec8ba2658b001ea2e16f10a8" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.271925 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9l4gg" event={"ID":"5046d57b-4ce3-4f2c-8cbc-0dee5590b458","Type":"ContainerDied","Data":"6f83ca5c19293c9af6936dd11315dbad3862b25157a23c26df1fa33c4f566212"} Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.271997 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f83ca5c19293c9af6936dd11315dbad3862b25157a23c26df1fa33c4f566212" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.273716 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-eeb1-account-create-update-hg4zn" event={"ID":"5872f688-95bb-4443-ba2d-f55ca9b05e97","Type":"ContainerDied","Data":"f865679f8bc3c5925e6771a6dc52e3a484e5bc23f0b92727148227cad163cf17"} Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.273758 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f865679f8bc3c5925e6771a6dc52e3a484e5bc23f0b92727148227cad163cf17" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.407263 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.411049 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.507902 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/5872f688-95bb-4443-ba2d-f55ca9b05e97-kube-api-access-9ltzp\") pod \"5872f688-95bb-4443-ba2d-f55ca9b05e97\" (UID: \"5872f688-95bb-4443-ba2d-f55ca9b05e97\") " Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.507967 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thswv\" (UniqueName: \"kubernetes.io/projected/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-kube-api-access-thswv\") pod \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\" (UID: \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\") " Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.508011 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5872f688-95bb-4443-ba2d-f55ca9b05e97-operator-scripts\") pod \"5872f688-95bb-4443-ba2d-f55ca9b05e97\" (UID: \"5872f688-95bb-4443-ba2d-f55ca9b05e97\") " Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.508048 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-operator-scripts\") pod \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\" (UID: \"5046d57b-4ce3-4f2c-8cbc-0dee5590b458\") " Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.508853 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5046d57b-4ce3-4f2c-8cbc-0dee5590b458" (UID: "5046d57b-4ce3-4f2c-8cbc-0dee5590b458"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.508912 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5872f688-95bb-4443-ba2d-f55ca9b05e97-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5872f688-95bb-4443-ba2d-f55ca9b05e97" (UID: "5872f688-95bb-4443-ba2d-f55ca9b05e97"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.512119 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5872f688-95bb-4443-ba2d-f55ca9b05e97-kube-api-access-9ltzp" (OuterVolumeSpecName: "kube-api-access-9ltzp") pod "5872f688-95bb-4443-ba2d-f55ca9b05e97" (UID: "5872f688-95bb-4443-ba2d-f55ca9b05e97"). InnerVolumeSpecName "kube-api-access-9ltzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.514742 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-kube-api-access-thswv" (OuterVolumeSpecName: "kube-api-access-thswv") pod "5046d57b-4ce3-4f2c-8cbc-0dee5590b458" (UID: "5046d57b-4ce3-4f2c-8cbc-0dee5590b458"). InnerVolumeSpecName "kube-api-access-thswv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.611430 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ltzp\" (UniqueName: \"kubernetes.io/projected/5872f688-95bb-4443-ba2d-f55ca9b05e97-kube-api-access-9ltzp\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.611462 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thswv\" (UniqueName: \"kubernetes.io/projected/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-kube-api-access-thswv\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.611475 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5872f688-95bb-4443-ba2d-f55ca9b05e97-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:16 crc kubenswrapper[4715]: I1203 22:15:16.611486 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5046d57b-4ce3-4f2c-8cbc-0dee5590b458-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:17 crc kubenswrapper[4715]: I1203 22:15:17.291707 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-eeb1-account-create-update-hg4zn" Dec 03 22:15:17 crc kubenswrapper[4715]: I1203 22:15:17.291715 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7f2gr" event={"ID":"18eb44f1-da35-4a10-b826-03c94f6b525e","Type":"ContainerStarted","Data":"c45699af7ab2664cb144dc48dab54aa1e6fb3fa8b569d6ea297af6b26ffc0d87"} Dec 03 22:15:17 crc kubenswrapper[4715]: I1203 22:15:17.291860 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9l4gg" Dec 03 22:15:17 crc kubenswrapper[4715]: I1203 22:15:17.326418 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-7f2gr" podStartSLOduration=2.797143821 podStartE2EDuration="11.326399093s" podCreationTimestamp="2025-12-03 22:15:06 +0000 UTC" firstStartedPulling="2025-12-03 22:15:07.71822468 +0000 UTC m=+2004.460935285" lastFinishedPulling="2025-12-03 22:15:16.247479952 +0000 UTC m=+2012.990190557" observedRunningTime="2025-12-03 22:15:17.324547124 +0000 UTC m=+2014.067257749" watchObservedRunningTime="2025-12-03 22:15:17.326399093 +0000 UTC m=+2014.069109688" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.581958 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-8cfqn"] Dec 03 22:15:18 crc kubenswrapper[4715]: E1203 22:15:18.582284 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5872f688-95bb-4443-ba2d-f55ca9b05e97" containerName="mariadb-account-create-update" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.582297 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5872f688-95bb-4443-ba2d-f55ca9b05e97" containerName="mariadb-account-create-update" Dec 03 22:15:18 crc kubenswrapper[4715]: E1203 22:15:18.582315 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerName="init" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.582321 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerName="init" Dec 03 22:15:18 crc kubenswrapper[4715]: E1203 22:15:18.582333 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerName="dnsmasq-dns" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.582339 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerName="dnsmasq-dns" Dec 03 22:15:18 crc kubenswrapper[4715]: E1203 22:15:18.582366 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5046d57b-4ce3-4f2c-8cbc-0dee5590b458" containerName="mariadb-database-create" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.582373 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5046d57b-4ce3-4f2c-8cbc-0dee5590b458" containerName="mariadb-database-create" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.582554 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e99e0a7-7ad8-4700-a5d1-000b78a2d03d" containerName="dnsmasq-dns" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.582580 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5046d57b-4ce3-4f2c-8cbc-0dee5590b458" containerName="mariadb-database-create" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.582588 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5872f688-95bb-4443-ba2d-f55ca9b05e97" containerName="mariadb-account-create-update" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.583076 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.585634 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.590191 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qcrqd" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.601412 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8cfqn"] Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.754655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-db-sync-config-data\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.754696 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-combined-ca-bundle\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.754742 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-config-data\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.755485 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkksz\" (UniqueName: \"kubernetes.io/projected/423716d6-494f-429a-98b4-696e1387bbc3-kube-api-access-pkksz\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.858219 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkksz\" (UniqueName: \"kubernetes.io/projected/423716d6-494f-429a-98b4-696e1387bbc3-kube-api-access-pkksz\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.858359 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-db-sync-config-data\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.858384 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-combined-ca-bundle\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.858426 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-config-data\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.863838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-config-data\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.868050 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-combined-ca-bundle\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.868437 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-db-sync-config-data\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.877115 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkksz\" (UniqueName: \"kubernetes.io/projected/423716d6-494f-429a-98b4-696e1387bbc3-kube-api-access-pkksz\") pod \"glance-db-sync-8cfqn\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:18 crc kubenswrapper[4715]: I1203 22:15:18.907843 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:19 crc kubenswrapper[4715]: I1203 22:15:19.329308 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8cfqn"] Dec 03 22:15:19 crc kubenswrapper[4715]: W1203 22:15:19.335929 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod423716d6_494f_429a_98b4_696e1387bbc3.slice/crio-d992fdd0acb7ad00ee2f0079e14b74982447a1a06d4f55ce15afcd3097be7397 WatchSource:0}: Error finding container d992fdd0acb7ad00ee2f0079e14b74982447a1a06d4f55ce15afcd3097be7397: Status 404 returned error can't find the container with id d992fdd0acb7ad00ee2f0079e14b74982447a1a06d4f55ce15afcd3097be7397 Dec 03 22:15:20 crc kubenswrapper[4715]: I1203 22:15:20.320785 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8cfqn" event={"ID":"423716d6-494f-429a-98b4-696e1387bbc3","Type":"ContainerStarted","Data":"d992fdd0acb7ad00ee2f0079e14b74982447a1a06d4f55ce15afcd3097be7397"} Dec 03 22:15:20 crc kubenswrapper[4715]: I1203 22:15:20.463728 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:15:20 crc kubenswrapper[4715]: I1203 22:15:20.527945 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vq29"] Dec 03 22:15:20 crc kubenswrapper[4715]: I1203 22:15:20.528323 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" podUID="28d0a607-f5ba-4e00-9e3d-111f0132026e" containerName="dnsmasq-dns" containerID="cri-o://c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b" gracePeriod=10 Dec 03 22:15:20 crc kubenswrapper[4715]: I1203 22:15:20.968475 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.102130 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg7nn\" (UniqueName: \"kubernetes.io/projected/28d0a607-f5ba-4e00-9e3d-111f0132026e-kube-api-access-bg7nn\") pod \"28d0a607-f5ba-4e00-9e3d-111f0132026e\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.102177 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-dns-svc\") pod \"28d0a607-f5ba-4e00-9e3d-111f0132026e\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.102264 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-config\") pod \"28d0a607-f5ba-4e00-9e3d-111f0132026e\" (UID: \"28d0a607-f5ba-4e00-9e3d-111f0132026e\") " Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.107820 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28d0a607-f5ba-4e00-9e3d-111f0132026e-kube-api-access-bg7nn" (OuterVolumeSpecName: "kube-api-access-bg7nn") pod "28d0a607-f5ba-4e00-9e3d-111f0132026e" (UID: "28d0a607-f5ba-4e00-9e3d-111f0132026e"). InnerVolumeSpecName "kube-api-access-bg7nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.143036 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "28d0a607-f5ba-4e00-9e3d-111f0132026e" (UID: "28d0a607-f5ba-4e00-9e3d-111f0132026e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.152118 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-config" (OuterVolumeSpecName: "config") pod "28d0a607-f5ba-4e00-9e3d-111f0132026e" (UID: "28d0a607-f5ba-4e00-9e3d-111f0132026e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.205164 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg7nn\" (UniqueName: \"kubernetes.io/projected/28d0a607-f5ba-4e00-9e3d-111f0132026e-kube-api-access-bg7nn\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.205200 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.205213 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28d0a607-f5ba-4e00-9e3d-111f0132026e-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.384145 4715 generic.go:334] "Generic (PLEG): container finished" podID="28d0a607-f5ba-4e00-9e3d-111f0132026e" containerID="c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b" exitCode=0 Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.384237 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" event={"ID":"28d0a607-f5ba-4e00-9e3d-111f0132026e","Type":"ContainerDied","Data":"c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b"} Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.384693 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" event={"ID":"28d0a607-f5ba-4e00-9e3d-111f0132026e","Type":"ContainerDied","Data":"008cdfd19e1ed9d5285c7b0971e7e1aea81047736d747d70f601e79971cae7bf"} Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.384727 4715 scope.go:117] "RemoveContainer" containerID="c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.384974 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vq29" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.407651 4715 scope.go:117] "RemoveContainer" containerID="131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.423794 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vq29"] Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.430170 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vq29"] Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.448313 4715 scope.go:117] "RemoveContainer" containerID="c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b" Dec 03 22:15:21 crc kubenswrapper[4715]: E1203 22:15:21.448744 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b\": container with ID starting with c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b not found: ID does not exist" containerID="c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.448776 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b"} err="failed to get container status \"c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b\": rpc error: code = NotFound desc = could not find container \"c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b\": container with ID starting with c25386ed5b3d4d70013c6b80b070e28c6d31e02eee30b56e726a7cff6bfbfa1b not found: ID does not exist" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.448797 4715 scope.go:117] "RemoveContainer" containerID="131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b" Dec 03 22:15:21 crc kubenswrapper[4715]: E1203 22:15:21.449140 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b\": container with ID starting with 131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b not found: ID does not exist" containerID="131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.449162 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b"} err="failed to get container status \"131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b\": rpc error: code = NotFound desc = could not find container \"131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b\": container with ID starting with 131d901cf41328cc996b9fa438fccb9de238eeb83fb11838b500123621da898b not found: ID does not exist" Dec 03 22:15:21 crc kubenswrapper[4715]: I1203 22:15:21.651557 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28d0a607-f5ba-4e00-9e3d-111f0132026e" path="/var/lib/kubelet/pods/28d0a607-f5ba-4e00-9e3d-111f0132026e/volumes" Dec 03 22:15:22 crc kubenswrapper[4715]: I1203 22:15:22.322152 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:22 crc kubenswrapper[4715]: E1203 22:15:22.322384 4715 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 22:15:22 crc kubenswrapper[4715]: E1203 22:15:22.322416 4715 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 22:15:22 crc kubenswrapper[4715]: E1203 22:15:22.322480 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift podName:d58c9eac-e46a-4a54-b1eb-27246d9c064e nodeName:}" failed. No retries permitted until 2025-12-03 22:15:38.322459293 +0000 UTC m=+2035.065169888 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift") pod "swift-storage-0" (UID: "d58c9eac-e46a-4a54-b1eb-27246d9c064e") : configmap "swift-ring-files" not found Dec 03 22:15:24 crc kubenswrapper[4715]: I1203 22:15:24.009480 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5fzvj" podUID="893e6b97-27e2-4372-8fc6-90562f2781bb" containerName="ovn-controller" probeResult="failure" output=< Dec 03 22:15:24 crc kubenswrapper[4715]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 22:15:24 crc kubenswrapper[4715]: > Dec 03 22:15:24 crc kubenswrapper[4715]: I1203 22:15:24.437320 4715 generic.go:334] "Generic (PLEG): container finished" podID="18eb44f1-da35-4a10-b826-03c94f6b525e" containerID="c45699af7ab2664cb144dc48dab54aa1e6fb3fa8b569d6ea297af6b26ffc0d87" exitCode=0 Dec 03 22:15:24 crc kubenswrapper[4715]: I1203 22:15:24.437388 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7f2gr" event={"ID":"18eb44f1-da35-4a10-b826-03c94f6b525e","Type":"ContainerDied","Data":"c45699af7ab2664cb144dc48dab54aa1e6fb3fa8b569d6ea297af6b26ffc0d87"} Dec 03 22:15:26 crc kubenswrapper[4715]: I1203 22:15:26.479439 4715 generic.go:334] "Generic (PLEG): container finished" podID="d6205840-7ea5-4060-844a-f6ce9feed200" containerID="f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd" exitCode=0 Dec 03 22:15:26 crc kubenswrapper[4715]: I1203 22:15:26.479563 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d6205840-7ea5-4060-844a-f6ce9feed200","Type":"ContainerDied","Data":"f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd"} Dec 03 22:15:26 crc kubenswrapper[4715]: I1203 22:15:26.482361 4715 generic.go:334] "Generic (PLEG): container finished" podID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerID="56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019" exitCode=0 Dec 03 22:15:26 crc kubenswrapper[4715]: I1203 22:15:26.482402 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5e2a0d28-27e4-4a86-8bf7-d5007af5162c","Type":"ContainerDied","Data":"56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019"} Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.023032 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-5fzvj" podUID="893e6b97-27e2-4372-8fc6-90562f2781bb" containerName="ovn-controller" probeResult="failure" output=< Dec 03 22:15:29 crc kubenswrapper[4715]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 22:15:29 crc kubenswrapper[4715]: > Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.033761 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.054842 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xq7sg" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.279795 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-5fzvj-config-4wvd6"] Dec 03 22:15:29 crc kubenswrapper[4715]: E1203 22:15:29.280177 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d0a607-f5ba-4e00-9e3d-111f0132026e" containerName="dnsmasq-dns" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.280190 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d0a607-f5ba-4e00-9e3d-111f0132026e" containerName="dnsmasq-dns" Dec 03 22:15:29 crc kubenswrapper[4715]: E1203 22:15:29.280211 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d0a607-f5ba-4e00-9e3d-111f0132026e" containerName="init" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.280216 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d0a607-f5ba-4e00-9e3d-111f0132026e" containerName="init" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.280366 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d0a607-f5ba-4e00-9e3d-111f0132026e" containerName="dnsmasq-dns" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.280972 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.283321 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.302122 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run-ovn\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.302210 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-additional-scripts\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.302271 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.302324 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-log-ovn\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.302344 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-scripts\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.302420 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcpm5\" (UniqueName: \"kubernetes.io/projected/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-kube-api-access-qcpm5\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.319726 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5fzvj-config-4wvd6"] Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.403718 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-additional-scripts\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.403798 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.403845 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-log-ovn\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.403866 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-scripts\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.403896 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcpm5\" (UniqueName: \"kubernetes.io/projected/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-kube-api-access-qcpm5\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.403925 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run-ovn\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.404193 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.404229 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run-ovn\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.404261 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-log-ovn\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.404702 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-additional-scripts\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.406306 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-scripts\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.428461 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcpm5\" (UniqueName: \"kubernetes.io/projected/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-kube-api-access-qcpm5\") pod \"ovn-controller-5fzvj-config-4wvd6\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:29 crc kubenswrapper[4715]: I1203 22:15:29.608582 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.446044 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.534375 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-7f2gr" event={"ID":"18eb44f1-da35-4a10-b826-03c94f6b525e","Type":"ContainerDied","Data":"cbcaf88329e50ad9332759d53a9236a069439aaa7dfdab1af3c19bd57252aaaa"} Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.534419 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbcaf88329e50ad9332759d53a9236a069439aaa7dfdab1af3c19bd57252aaaa" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.534499 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-7f2gr" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.648019 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-ring-data-devices\") pod \"18eb44f1-da35-4a10-b826-03c94f6b525e\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.648095 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/18eb44f1-da35-4a10-b826-03c94f6b525e-etc-swift\") pod \"18eb44f1-da35-4a10-b826-03c94f6b525e\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.648737 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-dispersionconf\") pod \"18eb44f1-da35-4a10-b826-03c94f6b525e\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.648847 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-258tq\" (UniqueName: \"kubernetes.io/projected/18eb44f1-da35-4a10-b826-03c94f6b525e-kube-api-access-258tq\") pod \"18eb44f1-da35-4a10-b826-03c94f6b525e\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.648971 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-swiftconf\") pod \"18eb44f1-da35-4a10-b826-03c94f6b525e\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.649030 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-scripts\") pod \"18eb44f1-da35-4a10-b826-03c94f6b525e\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.648964 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "18eb44f1-da35-4a10-b826-03c94f6b525e" (UID: "18eb44f1-da35-4a10-b826-03c94f6b525e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.649104 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-combined-ca-bundle\") pod \"18eb44f1-da35-4a10-b826-03c94f6b525e\" (UID: \"18eb44f1-da35-4a10-b826-03c94f6b525e\") " Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.649086 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18eb44f1-da35-4a10-b826-03c94f6b525e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "18eb44f1-da35-4a10-b826-03c94f6b525e" (UID: "18eb44f1-da35-4a10-b826-03c94f6b525e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.650207 4715 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/18eb44f1-da35-4a10-b826-03c94f6b525e-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.650232 4715 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.654486 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18eb44f1-da35-4a10-b826-03c94f6b525e-kube-api-access-258tq" (OuterVolumeSpecName: "kube-api-access-258tq") pod "18eb44f1-da35-4a10-b826-03c94f6b525e" (UID: "18eb44f1-da35-4a10-b826-03c94f6b525e"). InnerVolumeSpecName "kube-api-access-258tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.659446 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "18eb44f1-da35-4a10-b826-03c94f6b525e" (UID: "18eb44f1-da35-4a10-b826-03c94f6b525e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.675108 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18eb44f1-da35-4a10-b826-03c94f6b525e" (UID: "18eb44f1-da35-4a10-b826-03c94f6b525e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.676833 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "18eb44f1-da35-4a10-b826-03c94f6b525e" (UID: "18eb44f1-da35-4a10-b826-03c94f6b525e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.680624 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-scripts" (OuterVolumeSpecName: "scripts") pod "18eb44f1-da35-4a10-b826-03c94f6b525e" (UID: "18eb44f1-da35-4a10-b826-03c94f6b525e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.751824 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-258tq\" (UniqueName: \"kubernetes.io/projected/18eb44f1-da35-4a10-b826-03c94f6b525e-kube-api-access-258tq\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.751860 4715 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.751870 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18eb44f1-da35-4a10-b826-03c94f6b525e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.751882 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.751891 4715 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/18eb44f1-da35-4a10-b826-03c94f6b525e-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:31 crc kubenswrapper[4715]: I1203 22:15:31.837315 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-5fzvj-config-4wvd6"] Dec 03 22:15:31 crc kubenswrapper[4715]: W1203 22:15:31.850174 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb8a9670_8baf_4cc7_8dd5_9f97cc305fbb.slice/crio-788b5a77b3ab33230b42d7f695555934373565d535ada74fcaaa8131587e55f5 WatchSource:0}: Error finding container 788b5a77b3ab33230b42d7f695555934373565d535ada74fcaaa8131587e55f5: Status 404 returned error can't find the container with id 788b5a77b3ab33230b42d7f695555934373565d535ada74fcaaa8131587e55f5 Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.554291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5e2a0d28-27e4-4a86-8bf7-d5007af5162c","Type":"ContainerStarted","Data":"9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7"} Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.555188 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.555928 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8cfqn" event={"ID":"423716d6-494f-429a-98b4-696e1387bbc3","Type":"ContainerStarted","Data":"185c002ca496dccc6f7008715543c806b0ab76d68547ba111e9dbd24de2cf8cd"} Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.559221 4715 generic.go:334] "Generic (PLEG): container finished" podID="bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" containerID="b98f9d2ae2e384b9468509f35809ef612a35199ef9b1d08ad4f67bb0812d635f" exitCode=0 Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.559592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5fzvj-config-4wvd6" event={"ID":"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb","Type":"ContainerDied","Data":"b98f9d2ae2e384b9468509f35809ef612a35199ef9b1d08ad4f67bb0812d635f"} Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.559635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5fzvj-config-4wvd6" event={"ID":"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb","Type":"ContainerStarted","Data":"788b5a77b3ab33230b42d7f695555934373565d535ada74fcaaa8131587e55f5"} Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.562146 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d6205840-7ea5-4060-844a-f6ce9feed200","Type":"ContainerStarted","Data":"5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e"} Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.565528 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.593783 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=58.258894499 podStartE2EDuration="1m5.593749161s" podCreationTimestamp="2025-12-03 22:14:27 +0000 UTC" firstStartedPulling="2025-12-03 22:14:43.632726806 +0000 UTC m=+1980.375437411" lastFinishedPulling="2025-12-03 22:14:50.967581468 +0000 UTC m=+1987.710292073" observedRunningTime="2025-12-03 22:15:32.581443102 +0000 UTC m=+2029.324153697" watchObservedRunningTime="2025-12-03 22:15:32.593749161 +0000 UTC m=+2029.336459786" Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.619030 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-8cfqn" podStartSLOduration=2.605611592 podStartE2EDuration="14.619005556s" podCreationTimestamp="2025-12-03 22:15:18 +0000 UTC" firstStartedPulling="2025-12-03 22:15:19.33816829 +0000 UTC m=+2016.080878885" lastFinishedPulling="2025-12-03 22:15:31.351562254 +0000 UTC m=+2028.094272849" observedRunningTime="2025-12-03 22:15:32.614052644 +0000 UTC m=+2029.356763279" watchObservedRunningTime="2025-12-03 22:15:32.619005556 +0000 UTC m=+2029.361716161" Dec 03 22:15:32 crc kubenswrapper[4715]: I1203 22:15:32.647388 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=57.198999367 podStartE2EDuration="1m4.647368675s" podCreationTimestamp="2025-12-03 22:14:28 +0000 UTC" firstStartedPulling="2025-12-03 22:14:44.244483415 +0000 UTC m=+1980.987194030" lastFinishedPulling="2025-12-03 22:14:51.692852743 +0000 UTC m=+1988.435563338" observedRunningTime="2025-12-03 22:15:32.64010694 +0000 UTC m=+2029.382817535" watchObservedRunningTime="2025-12-03 22:15:32.647368675 +0000 UTC m=+2029.390079270" Dec 03 22:15:33 crc kubenswrapper[4715]: I1203 22:15:33.951581 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.015831 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-5fzvj" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.101759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run\") pod \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.101897 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-log-ovn\") pod \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.101928 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run" (OuterVolumeSpecName: "var-run") pod "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" (UID: "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.101952 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-additional-scripts\") pod \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.102000 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" (UID: "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.102154 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcpm5\" (UniqueName: \"kubernetes.io/projected/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-kube-api-access-qcpm5\") pod \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.102213 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run-ovn\") pod \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.102276 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-scripts\") pod \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\" (UID: \"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb\") " Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.102397 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" (UID: "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.102677 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" (UID: "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.103184 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-scripts" (OuterVolumeSpecName: "scripts") pod "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" (UID: "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.103214 4715 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.103240 4715 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.103282 4715 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.103299 4715 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.109653 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-kube-api-access-qcpm5" (OuterVolumeSpecName: "kube-api-access-qcpm5") pod "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" (UID: "bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb"). InnerVolumeSpecName "kube-api-access-qcpm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.205805 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcpm5\" (UniqueName: \"kubernetes.io/projected/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-kube-api-access-qcpm5\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.205852 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.586294 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-5fzvj-config-4wvd6" event={"ID":"bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb","Type":"ContainerDied","Data":"788b5a77b3ab33230b42d7f695555934373565d535ada74fcaaa8131587e55f5"} Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.586330 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-5fzvj-config-4wvd6" Dec 03 22:15:34 crc kubenswrapper[4715]: I1203 22:15:34.586349 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="788b5a77b3ab33230b42d7f695555934373565d535ada74fcaaa8131587e55f5" Dec 03 22:15:35 crc kubenswrapper[4715]: I1203 22:15:35.064990 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-5fzvj-config-4wvd6"] Dec 03 22:15:35 crc kubenswrapper[4715]: I1203 22:15:35.076656 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-5fzvj-config-4wvd6"] Dec 03 22:15:35 crc kubenswrapper[4715]: I1203 22:15:35.160020 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:15:35 crc kubenswrapper[4715]: I1203 22:15:35.160077 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:15:35 crc kubenswrapper[4715]: I1203 22:15:35.657495 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" path="/var/lib/kubelet/pods/bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb/volumes" Dec 03 22:15:38 crc kubenswrapper[4715]: I1203 22:15:38.381659 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:38 crc kubenswrapper[4715]: I1203 22:15:38.392248 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/d58c9eac-e46a-4a54-b1eb-27246d9c064e-etc-swift\") pod \"swift-storage-0\" (UID: \"d58c9eac-e46a-4a54-b1eb-27246d9c064e\") " pod="openstack/swift-storage-0" Dec 03 22:15:38 crc kubenswrapper[4715]: I1203 22:15:38.447342 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 22:15:38 crc kubenswrapper[4715]: I1203 22:15:38.636538 4715 generic.go:334] "Generic (PLEG): container finished" podID="423716d6-494f-429a-98b4-696e1387bbc3" containerID="185c002ca496dccc6f7008715543c806b0ab76d68547ba111e9dbd24de2cf8cd" exitCode=0 Dec 03 22:15:38 crc kubenswrapper[4715]: I1203 22:15:38.636606 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8cfqn" event={"ID":"423716d6-494f-429a-98b4-696e1387bbc3","Type":"ContainerDied","Data":"185c002ca496dccc6f7008715543c806b0ab76d68547ba111e9dbd24de2cf8cd"} Dec 03 22:15:38 crc kubenswrapper[4715]: I1203 22:15:38.853182 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 22:15:39 crc kubenswrapper[4715]: I1203 22:15:39.654307 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"450b27e616272a8c5b4a15ff79dd1518eedf1c555159ee635188f9510ce48fb4"} Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.047537 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.112979 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-config-data\") pod \"423716d6-494f-429a-98b4-696e1387bbc3\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.113112 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-combined-ca-bundle\") pod \"423716d6-494f-429a-98b4-696e1387bbc3\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.113138 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-db-sync-config-data\") pod \"423716d6-494f-429a-98b4-696e1387bbc3\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.113218 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkksz\" (UniqueName: \"kubernetes.io/projected/423716d6-494f-429a-98b4-696e1387bbc3-kube-api-access-pkksz\") pod \"423716d6-494f-429a-98b4-696e1387bbc3\" (UID: \"423716d6-494f-429a-98b4-696e1387bbc3\") " Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.119924 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/423716d6-494f-429a-98b4-696e1387bbc3-kube-api-access-pkksz" (OuterVolumeSpecName: "kube-api-access-pkksz") pod "423716d6-494f-429a-98b4-696e1387bbc3" (UID: "423716d6-494f-429a-98b4-696e1387bbc3"). InnerVolumeSpecName "kube-api-access-pkksz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.121231 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "423716d6-494f-429a-98b4-696e1387bbc3" (UID: "423716d6-494f-429a-98b4-696e1387bbc3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.163943 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "423716d6-494f-429a-98b4-696e1387bbc3" (UID: "423716d6-494f-429a-98b4-696e1387bbc3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.170001 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-config-data" (OuterVolumeSpecName: "config-data") pod "423716d6-494f-429a-98b4-696e1387bbc3" (UID: "423716d6-494f-429a-98b4-696e1387bbc3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.215004 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.215031 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.215041 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/423716d6-494f-429a-98b4-696e1387bbc3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.215051 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkksz\" (UniqueName: \"kubernetes.io/projected/423716d6-494f-429a-98b4-696e1387bbc3-kube-api-access-pkksz\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.666597 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8cfqn" event={"ID":"423716d6-494f-429a-98b4-696e1387bbc3","Type":"ContainerDied","Data":"d992fdd0acb7ad00ee2f0079e14b74982447a1a06d4f55ce15afcd3097be7397"} Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.666649 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d992fdd0acb7ad00ee2f0079e14b74982447a1a06d4f55ce15afcd3097be7397" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.669676 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8cfqn" Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.681548 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"4aaaf1490f6da72f6b5f46dd7c132300d514ea84e52528bfdad45a1d37c8bb83"} Dec 03 22:15:40 crc kubenswrapper[4715]: I1203 22:15:40.681633 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"c00638b9d130bbacf53a24f8640bfaaa32623585a017c16a54b704916a304b8e"} Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.025132 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-dcq57"] Dec 03 22:15:41 crc kubenswrapper[4715]: E1203 22:15:41.025477 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18eb44f1-da35-4a10-b826-03c94f6b525e" containerName="swift-ring-rebalance" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.025495 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="18eb44f1-da35-4a10-b826-03c94f6b525e" containerName="swift-ring-rebalance" Dec 03 22:15:41 crc kubenswrapper[4715]: E1203 22:15:41.025516 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" containerName="ovn-config" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.025524 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" containerName="ovn-config" Dec 03 22:15:41 crc kubenswrapper[4715]: E1203 22:15:41.025552 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="423716d6-494f-429a-98b4-696e1387bbc3" containerName="glance-db-sync" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.025561 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="423716d6-494f-429a-98b4-696e1387bbc3" containerName="glance-db-sync" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.025719 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="18eb44f1-da35-4a10-b826-03c94f6b525e" containerName="swift-ring-rebalance" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.025753 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb8a9670-8baf-4cc7-8dd5-9f97cc305fbb" containerName="ovn-config" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.025766 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="423716d6-494f-429a-98b4-696e1387bbc3" containerName="glance-db-sync" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.027279 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.049202 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-dcq57"] Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.135707 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvjsz\" (UniqueName: \"kubernetes.io/projected/c735ae57-52d7-4158-a5f1-31c7294828aa-kube-api-access-dvjsz\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.135779 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.135874 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-dns-svc\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.135892 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.135925 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-config\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.237471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvjsz\" (UniqueName: \"kubernetes.io/projected/c735ae57-52d7-4158-a5f1-31c7294828aa-kube-api-access-dvjsz\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.237965 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.238038 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-dns-svc\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.238059 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.238086 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-config\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.238976 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-config\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.239022 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-dns-svc\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.239171 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.239245 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.271472 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvjsz\" (UniqueName: \"kubernetes.io/projected/c735ae57-52d7-4158-a5f1-31c7294828aa-kube-api-access-dvjsz\") pod \"dnsmasq-dns-74dc88fc-dcq57\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.342871 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.692477 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"61d275ebd66b64855bac89243f690ece4f207643dd8b8f53027f30fb19cd34ee"} Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.692786 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"8d487f7d9db55581035b0fd2f6bee52bccf3ea3dc8ee6ba0aa3dea1fc7f55c32"} Dec 03 22:15:41 crc kubenswrapper[4715]: I1203 22:15:41.805152 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-dcq57"] Dec 03 22:15:42 crc kubenswrapper[4715]: W1203 22:15:42.042336 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc735ae57_52d7_4158_a5f1_31c7294828aa.slice/crio-ae42694c99203d6e89c32416fe93e5a0b1e52c146e485675cb2ca5a4bd6a3641 WatchSource:0}: Error finding container ae42694c99203d6e89c32416fe93e5a0b1e52c146e485675cb2ca5a4bd6a3641: Status 404 returned error can't find the container with id ae42694c99203d6e89c32416fe93e5a0b1e52c146e485675cb2ca5a4bd6a3641 Dec 03 22:15:42 crc kubenswrapper[4715]: I1203 22:15:42.701941 4715 generic.go:334] "Generic (PLEG): container finished" podID="c735ae57-52d7-4158-a5f1-31c7294828aa" containerID="ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71" exitCode=0 Dec 03 22:15:42 crc kubenswrapper[4715]: I1203 22:15:42.702149 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" event={"ID":"c735ae57-52d7-4158-a5f1-31c7294828aa","Type":"ContainerDied","Data":"ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71"} Dec 03 22:15:42 crc kubenswrapper[4715]: I1203 22:15:42.702436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" event={"ID":"c735ae57-52d7-4158-a5f1-31c7294828aa","Type":"ContainerStarted","Data":"ae42694c99203d6e89c32416fe93e5a0b1e52c146e485675cb2ca5a4bd6a3641"} Dec 03 22:15:42 crc kubenswrapper[4715]: I1203 22:15:42.710475 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"a07ed8c7602fe306280610e7b73c200a6657dcde1666af691d4370b0a9f8f8cf"} Dec 03 22:15:42 crc kubenswrapper[4715]: I1203 22:15:42.710818 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"4a33fd5fb6b634b10da22da13ab78fe4c61af8e3e83ede2a6e375e608f10a568"} Dec 03 22:15:42 crc kubenswrapper[4715]: I1203 22:15:42.710964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"b990a3e12b082695c047c063c4aac5d0d4889faf84b6dcb00e31444a8082867b"} Dec 03 22:15:43 crc kubenswrapper[4715]: I1203 22:15:43.719781 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" event={"ID":"c735ae57-52d7-4158-a5f1-31c7294828aa","Type":"ContainerStarted","Data":"1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66"} Dec 03 22:15:43 crc kubenswrapper[4715]: I1203 22:15:43.720316 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:43 crc kubenswrapper[4715]: I1203 22:15:43.726036 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"9eb7adc8b12cc5e288fcb4e0d79bebf2919744d891df2cdac9a8cea3a42d7745"} Dec 03 22:15:43 crc kubenswrapper[4715]: I1203 22:15:43.749366 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" podStartSLOduration=2.749346385 podStartE2EDuration="2.749346385s" podCreationTimestamp="2025-12-03 22:15:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:15:43.749310114 +0000 UTC m=+2040.492020709" watchObservedRunningTime="2025-12-03 22:15:43.749346385 +0000 UTC m=+2040.492056980" Dec 03 22:15:44 crc kubenswrapper[4715]: I1203 22:15:44.740123 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"b10318d3a230df9fd0d17aa63cce89c7088c90bcbd654cbf56f242ca97859f4d"} Dec 03 22:15:44 crc kubenswrapper[4715]: I1203 22:15:44.740587 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"f7cddedaad509e7d2e82bb728e743ee637e11bc5668de71c41c3edf61cf5b590"} Dec 03 22:15:45 crc kubenswrapper[4715]: I1203 22:15:45.764879 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"4dfd14ab7d9917224c38142c2dc772193c270599881f7c76732072cb9680b6f7"} Dec 03 22:15:45 crc kubenswrapper[4715]: I1203 22:15:45.765291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"17e6bb171b42e0f88b69d9ac16d1956b51f08d463e82494089a52f76bf385e2a"} Dec 03 22:15:45 crc kubenswrapper[4715]: I1203 22:15:45.765306 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"45d2fbdab93f5a4e69111dcd42b61ed50096e772e8dd4f824c99be5dc48547b9"} Dec 03 22:15:45 crc kubenswrapper[4715]: I1203 22:15:45.765320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"5cab663c7c2e5d6b02b445294fc77d5a4d94b3c93c583526707677ba466d401a"} Dec 03 22:15:45 crc kubenswrapper[4715]: I1203 22:15:45.765332 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"d58c9eac-e46a-4a54-b1eb-27246d9c064e","Type":"ContainerStarted","Data":"44d9d5a1123ba07fecc35cb1fb7db8c2f2b5702f4bac695d7bc03c463fbcdeef"} Dec 03 22:15:45 crc kubenswrapper[4715]: I1203 22:15:45.819200 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=35.554987514 podStartE2EDuration="40.819176974s" podCreationTimestamp="2025-12-03 22:15:05 +0000 UTC" firstStartedPulling="2025-12-03 22:15:38.848819459 +0000 UTC m=+2035.591530094" lastFinishedPulling="2025-12-03 22:15:44.113008959 +0000 UTC m=+2040.855719554" observedRunningTime="2025-12-03 22:15:45.812459384 +0000 UTC m=+2042.555169979" watchObservedRunningTime="2025-12-03 22:15:45.819176974 +0000 UTC m=+2042.561887569" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.194100 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-dcq57"] Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.194428 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" podUID="c735ae57-52d7-4158-a5f1-31c7294828aa" containerName="dnsmasq-dns" containerID="cri-o://1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66" gracePeriod=10 Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.231286 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-f5kgj"] Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.233081 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.236475 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.293332 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-f5kgj"] Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.390143 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.390480 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-config\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.390534 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.390560 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.390578 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.390601 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcxpp\" (UniqueName: \"kubernetes.io/projected/74cdaa65-5773-4ff8-b84d-31100df6f576-kube-api-access-dcxpp\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.492066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.492148 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-config\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.492196 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.492228 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.492259 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.492289 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcxpp\" (UniqueName: \"kubernetes.io/projected/74cdaa65-5773-4ff8-b84d-31100df6f576-kube-api-access-dcxpp\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.493078 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.493663 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.493681 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.493763 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-config\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.494373 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.522346 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcxpp\" (UniqueName: \"kubernetes.io/projected/74cdaa65-5773-4ff8-b84d-31100df6f576-kube-api-access-dcxpp\") pod \"dnsmasq-dns-5f59b8f679-f5kgj\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.659928 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.665403 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.775302 4715 generic.go:334] "Generic (PLEG): container finished" podID="c735ae57-52d7-4158-a5f1-31c7294828aa" containerID="1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66" exitCode=0 Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.775354 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.775406 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" event={"ID":"c735ae57-52d7-4158-a5f1-31c7294828aa","Type":"ContainerDied","Data":"1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66"} Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.775455 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-dcq57" event={"ID":"c735ae57-52d7-4158-a5f1-31c7294828aa","Type":"ContainerDied","Data":"ae42694c99203d6e89c32416fe93e5a0b1e52c146e485675cb2ca5a4bd6a3641"} Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.775477 4715 scope.go:117] "RemoveContainer" containerID="1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.797079 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-sb\") pod \"c735ae57-52d7-4158-a5f1-31c7294828aa\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.797414 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-config\") pod \"c735ae57-52d7-4158-a5f1-31c7294828aa\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.797441 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-nb\") pod \"c735ae57-52d7-4158-a5f1-31c7294828aa\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.797638 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-dns-svc\") pod \"c735ae57-52d7-4158-a5f1-31c7294828aa\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.797759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvjsz\" (UniqueName: \"kubernetes.io/projected/c735ae57-52d7-4158-a5f1-31c7294828aa-kube-api-access-dvjsz\") pod \"c735ae57-52d7-4158-a5f1-31c7294828aa\" (UID: \"c735ae57-52d7-4158-a5f1-31c7294828aa\") " Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.803355 4715 scope.go:117] "RemoveContainer" containerID="ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.804772 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c735ae57-52d7-4158-a5f1-31c7294828aa-kube-api-access-dvjsz" (OuterVolumeSpecName: "kube-api-access-dvjsz") pod "c735ae57-52d7-4158-a5f1-31c7294828aa" (UID: "c735ae57-52d7-4158-a5f1-31c7294828aa"). InnerVolumeSpecName "kube-api-access-dvjsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.841372 4715 scope.go:117] "RemoveContainer" containerID="1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66" Dec 03 22:15:46 crc kubenswrapper[4715]: E1203 22:15:46.844925 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66\": container with ID starting with 1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66 not found: ID does not exist" containerID="1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.844962 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66"} err="failed to get container status \"1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66\": rpc error: code = NotFound desc = could not find container \"1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66\": container with ID starting with 1e14adb60738ac7ee836c3281177e300b652fe7cd481baec02a793cf3c7d2f66 not found: ID does not exist" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.844999 4715 scope.go:117] "RemoveContainer" containerID="ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71" Dec 03 22:15:46 crc kubenswrapper[4715]: E1203 22:15:46.847973 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71\": container with ID starting with ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71 not found: ID does not exist" containerID="ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.848028 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71"} err="failed to get container status \"ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71\": rpc error: code = NotFound desc = could not find container \"ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71\": container with ID starting with ebc6857b0f7fcb24ba68fac90093d5f4efb2d65c9c582c063ad0814d65890d71 not found: ID does not exist" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.856361 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-config" (OuterVolumeSpecName: "config") pod "c735ae57-52d7-4158-a5f1-31c7294828aa" (UID: "c735ae57-52d7-4158-a5f1-31c7294828aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.872585 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c735ae57-52d7-4158-a5f1-31c7294828aa" (UID: "c735ae57-52d7-4158-a5f1-31c7294828aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.876605 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c735ae57-52d7-4158-a5f1-31c7294828aa" (UID: "c735ae57-52d7-4158-a5f1-31c7294828aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.878256 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c735ae57-52d7-4158-a5f1-31c7294828aa" (UID: "c735ae57-52d7-4158-a5f1-31c7294828aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.904341 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvjsz\" (UniqueName: \"kubernetes.io/projected/c735ae57-52d7-4158-a5f1-31c7294828aa-kube-api-access-dvjsz\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.904381 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.904397 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.904410 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:46 crc kubenswrapper[4715]: I1203 22:15:46.904427 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c735ae57-52d7-4158-a5f1-31c7294828aa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:47 crc kubenswrapper[4715]: I1203 22:15:47.060063 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-f5kgj"] Dec 03 22:15:47 crc kubenswrapper[4715]: I1203 22:15:47.144893 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-dcq57"] Dec 03 22:15:47 crc kubenswrapper[4715]: I1203 22:15:47.153474 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-dcq57"] Dec 03 22:15:47 crc kubenswrapper[4715]: I1203 22:15:47.648218 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c735ae57-52d7-4158-a5f1-31c7294828aa" path="/var/lib/kubelet/pods/c735ae57-52d7-4158-a5f1-31c7294828aa/volumes" Dec 03 22:15:47 crc kubenswrapper[4715]: I1203 22:15:47.782717 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" event={"ID":"74cdaa65-5773-4ff8-b84d-31100df6f576","Type":"ContainerStarted","Data":"3b4e150e07844647a418fca22372f5889863a016fbb1141e1d0f910e45d9b46a"} Dec 03 22:15:48 crc kubenswrapper[4715]: I1203 22:15:48.812738 4715 generic.go:334] "Generic (PLEG): container finished" podID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerID="906b471218ec478389c8bf75c00fd4fe476fbf74018532f45fc55547fc837028" exitCode=0 Dec 03 22:15:48 crc kubenswrapper[4715]: I1203 22:15:48.812796 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" event={"ID":"74cdaa65-5773-4ff8-b84d-31100df6f576","Type":"ContainerDied","Data":"906b471218ec478389c8bf75c00fd4fe476fbf74018532f45fc55547fc837028"} Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.365817 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.750669 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.774796 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-j4dtp"] Dec 03 22:15:49 crc kubenswrapper[4715]: E1203 22:15:49.775123 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c735ae57-52d7-4158-a5f1-31c7294828aa" containerName="dnsmasq-dns" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.775151 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c735ae57-52d7-4158-a5f1-31c7294828aa" containerName="dnsmasq-dns" Dec 03 22:15:49 crc kubenswrapper[4715]: E1203 22:15:49.775180 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c735ae57-52d7-4158-a5f1-31c7294828aa" containerName="init" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.775191 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c735ae57-52d7-4158-a5f1-31c7294828aa" containerName="init" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.775372 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c735ae57-52d7-4158-a5f1-31c7294828aa" containerName="dnsmasq-dns" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.775890 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.796616 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-j4dtp"] Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.840843 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" event={"ID":"74cdaa65-5773-4ff8-b84d-31100df6f576","Type":"ContainerStarted","Data":"c7755f1a98a28567ccae06ca7899178a25b184da6e2f67de8f9a4ce23cd2b684"} Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.840994 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.873729 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svbtq\" (UniqueName: \"kubernetes.io/projected/75b6412c-5008-4966-b1d9-55f9846d9cd1-kube-api-access-svbtq\") pod \"cinder-db-create-j4dtp\" (UID: \"75b6412c-5008-4966-b1d9-55f9846d9cd1\") " pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.873812 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b6412c-5008-4966-b1d9-55f9846d9cd1-operator-scripts\") pod \"cinder-db-create-j4dtp\" (UID: \"75b6412c-5008-4966-b1d9-55f9846d9cd1\") " pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.879619 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" podStartSLOduration=3.879596964 podStartE2EDuration="3.879596964s" podCreationTimestamp="2025-12-03 22:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:15:49.872782812 +0000 UTC m=+2046.615493407" watchObservedRunningTime="2025-12-03 22:15:49.879596964 +0000 UTC m=+2046.622307559" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.904846 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-tzppt"] Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.906000 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.916581 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-de87-account-create-update-4tn28"] Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.917890 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.921166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.926040 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-tzppt"] Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.935831 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-de87-account-create-update-4tn28"] Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.975057 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svbtq\" (UniqueName: \"kubernetes.io/projected/75b6412c-5008-4966-b1d9-55f9846d9cd1-kube-api-access-svbtq\") pod \"cinder-db-create-j4dtp\" (UID: \"75b6412c-5008-4966-b1d9-55f9846d9cd1\") " pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.975132 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b5688d-8ef1-44f9-b05e-60e99834e994-operator-scripts\") pod \"barbican-db-create-tzppt\" (UID: \"60b5688d-8ef1-44f9-b05e-60e99834e994\") " pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.975172 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl78h\" (UniqueName: \"kubernetes.io/projected/60b5688d-8ef1-44f9-b05e-60e99834e994-kube-api-access-dl78h\") pod \"barbican-db-create-tzppt\" (UID: \"60b5688d-8ef1-44f9-b05e-60e99834e994\") " pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.975194 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b6412c-5008-4966-b1d9-55f9846d9cd1-operator-scripts\") pod \"cinder-db-create-j4dtp\" (UID: \"75b6412c-5008-4966-b1d9-55f9846d9cd1\") " pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.975852 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b6412c-5008-4966-b1d9-55f9846d9cd1-operator-scripts\") pod \"cinder-db-create-j4dtp\" (UID: \"75b6412c-5008-4966-b1d9-55f9846d9cd1\") " pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.986287 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-fd69-account-create-update-n7kqs"] Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.987373 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.989243 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 22:15:49 crc kubenswrapper[4715]: I1203 22:15:49.992789 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fd69-account-create-update-n7kqs"] Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.019718 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svbtq\" (UniqueName: \"kubernetes.io/projected/75b6412c-5008-4966-b1d9-55f9846d9cd1-kube-api-access-svbtq\") pod \"cinder-db-create-j4dtp\" (UID: \"75b6412c-5008-4966-b1d9-55f9846d9cd1\") " pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.072216 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-hg7kd"] Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.073192 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.076780 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-operator-scripts\") pod \"barbican-fd69-account-create-update-n7kqs\" (UID: \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\") " pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.076828 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075e50c6-4819-42c7-88b3-a73a7f959f81-operator-scripts\") pod \"cinder-de87-account-create-update-4tn28\" (UID: \"075e50c6-4819-42c7-88b3-a73a7f959f81\") " pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.076872 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b5688d-8ef1-44f9-b05e-60e99834e994-operator-scripts\") pod \"barbican-db-create-tzppt\" (UID: \"60b5688d-8ef1-44f9-b05e-60e99834e994\") " pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.076902 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl78h\" (UniqueName: \"kubernetes.io/projected/60b5688d-8ef1-44f9-b05e-60e99834e994-kube-api-access-dl78h\") pod \"barbican-db-create-tzppt\" (UID: \"60b5688d-8ef1-44f9-b05e-60e99834e994\") " pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.076972 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p6h9\" (UniqueName: \"kubernetes.io/projected/075e50c6-4819-42c7-88b3-a73a7f959f81-kube-api-access-8p6h9\") pod \"cinder-de87-account-create-update-4tn28\" (UID: \"075e50c6-4819-42c7-88b3-a73a7f959f81\") " pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.076995 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42p9w\" (UniqueName: \"kubernetes.io/projected/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-kube-api-access-42p9w\") pod \"barbican-fd69-account-create-update-n7kqs\" (UID: \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\") " pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.077741 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b5688d-8ef1-44f9-b05e-60e99834e994-operator-scripts\") pod \"barbican-db-create-tzppt\" (UID: \"60b5688d-8ef1-44f9-b05e-60e99834e994\") " pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.081318 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hg7kd"] Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.117351 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.123644 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl78h\" (UniqueName: \"kubernetes.io/projected/60b5688d-8ef1-44f9-b05e-60e99834e994-kube-api-access-dl78h\") pod \"barbican-db-create-tzppt\" (UID: \"60b5688d-8ef1-44f9-b05e-60e99834e994\") " pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.178054 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8w8l\" (UniqueName: \"kubernetes.io/projected/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-kube-api-access-v8w8l\") pod \"neutron-db-create-hg7kd\" (UID: \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\") " pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.178152 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p6h9\" (UniqueName: \"kubernetes.io/projected/075e50c6-4819-42c7-88b3-a73a7f959f81-kube-api-access-8p6h9\") pod \"cinder-de87-account-create-update-4tn28\" (UID: \"075e50c6-4819-42c7-88b3-a73a7f959f81\") " pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.178181 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42p9w\" (UniqueName: \"kubernetes.io/projected/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-kube-api-access-42p9w\") pod \"barbican-fd69-account-create-update-n7kqs\" (UID: \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\") " pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.178218 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-operator-scripts\") pod \"barbican-fd69-account-create-update-n7kqs\" (UID: \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\") " pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.178258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-operator-scripts\") pod \"neutron-db-create-hg7kd\" (UID: \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\") " pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.178284 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075e50c6-4819-42c7-88b3-a73a7f959f81-operator-scripts\") pod \"cinder-de87-account-create-update-4tn28\" (UID: \"075e50c6-4819-42c7-88b3-a73a7f959f81\") " pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.179081 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-operator-scripts\") pod \"barbican-fd69-account-create-update-n7kqs\" (UID: \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\") " pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.179101 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075e50c6-4819-42c7-88b3-a73a7f959f81-operator-scripts\") pod \"cinder-de87-account-create-update-4tn28\" (UID: \"075e50c6-4819-42c7-88b3-a73a7f959f81\") " pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.197966 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42p9w\" (UniqueName: \"kubernetes.io/projected/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-kube-api-access-42p9w\") pod \"barbican-fd69-account-create-update-n7kqs\" (UID: \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\") " pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.198293 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p6h9\" (UniqueName: \"kubernetes.io/projected/075e50c6-4819-42c7-88b3-a73a7f959f81-kube-api-access-8p6h9\") pod \"cinder-de87-account-create-update-4tn28\" (UID: \"075e50c6-4819-42c7-88b3-a73a7f959f81\") " pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.224878 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.245080 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.282511 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-operator-scripts\") pod \"neutron-db-create-hg7kd\" (UID: \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\") " pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.282616 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8w8l\" (UniqueName: \"kubernetes.io/projected/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-kube-api-access-v8w8l\") pod \"neutron-db-create-hg7kd\" (UID: \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\") " pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.283587 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-operator-scripts\") pod \"neutron-db-create-hg7kd\" (UID: \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\") " pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.284362 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-d7jfj"] Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.292730 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.301411 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9bcdn" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.301655 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.302248 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.306446 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.308750 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.339337 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-d7jfj"] Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.395300 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-combined-ca-bundle\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.426049 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wsc7\" (UniqueName: \"kubernetes.io/projected/7e7668eb-91d3-4172-9b37-048835ae6899-kube-api-access-5wsc7\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.426151 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-config-data\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.394463 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8w8l\" (UniqueName: \"kubernetes.io/projected/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-kube-api-access-v8w8l\") pod \"neutron-db-create-hg7kd\" (UID: \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\") " pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.400779 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-e283-account-create-update-r7wc9"] Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.428401 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.444442 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.484038 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.493459 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e283-account-create-update-r7wc9"] Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.531450 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wsc7\" (UniqueName: \"kubernetes.io/projected/7e7668eb-91d3-4172-9b37-048835ae6899-kube-api-access-5wsc7\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.532079 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-config-data\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.532185 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-combined-ca-bundle\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.532287 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rvjb\" (UniqueName: \"kubernetes.io/projected/79758367-7636-4dcd-9026-b4008520c52d-kube-api-access-5rvjb\") pod \"neutron-e283-account-create-update-r7wc9\" (UID: \"79758367-7636-4dcd-9026-b4008520c52d\") " pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.532362 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79758367-7636-4dcd-9026-b4008520c52d-operator-scripts\") pod \"neutron-e283-account-create-update-r7wc9\" (UID: \"79758367-7636-4dcd-9026-b4008520c52d\") " pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.554188 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-combined-ca-bundle\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.554543 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-config-data\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.570997 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wsc7\" (UniqueName: \"kubernetes.io/projected/7e7668eb-91d3-4172-9b37-048835ae6899-kube-api-access-5wsc7\") pod \"keystone-db-sync-d7jfj\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.651981 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rvjb\" (UniqueName: \"kubernetes.io/projected/79758367-7636-4dcd-9026-b4008520c52d-kube-api-access-5rvjb\") pod \"neutron-e283-account-create-update-r7wc9\" (UID: \"79758367-7636-4dcd-9026-b4008520c52d\") " pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.652033 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79758367-7636-4dcd-9026-b4008520c52d-operator-scripts\") pod \"neutron-e283-account-create-update-r7wc9\" (UID: \"79758367-7636-4dcd-9026-b4008520c52d\") " pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.652768 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79758367-7636-4dcd-9026-b4008520c52d-operator-scripts\") pod \"neutron-e283-account-create-update-r7wc9\" (UID: \"79758367-7636-4dcd-9026-b4008520c52d\") " pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.671217 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-j4dtp"] Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.674278 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rvjb\" (UniqueName: \"kubernetes.io/projected/79758367-7636-4dcd-9026-b4008520c52d-kube-api-access-5rvjb\") pod \"neutron-e283-account-create-update-r7wc9\" (UID: \"79758367-7636-4dcd-9026-b4008520c52d\") " pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.699276 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.853883 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j4dtp" event={"ID":"75b6412c-5008-4966-b1d9-55f9846d9cd1","Type":"ContainerStarted","Data":"758aeb89be7cb6740d2a74aca847aa94d7755e164be384abe39872fef05977ad"} Dec 03 22:15:50 crc kubenswrapper[4715]: I1203 22:15:50.870947 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.171425 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-de87-account-create-update-4tn28"] Dec 03 22:15:51 crc kubenswrapper[4715]: W1203 22:15:51.173212 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2e31ed8_58e8_4ee3_a7d7_ee6f8dec6ab7.slice/crio-7c116a2f8f996690099acfe86b930445ded2bdb794dc5514f48b357034d2325e WatchSource:0}: Error finding container 7c116a2f8f996690099acfe86b930445ded2bdb794dc5514f48b357034d2325e: Status 404 returned error can't find the container with id 7c116a2f8f996690099acfe86b930445ded2bdb794dc5514f48b357034d2325e Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.180154 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fd69-account-create-update-n7kqs"] Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.189627 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-tzppt"] Dec 03 22:15:51 crc kubenswrapper[4715]: W1203 22:15:51.255143 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod075e50c6_4819_42c7_88b3_a73a7f959f81.slice/crio-03fb71018b6820b54cc186e02849883a08777f4ab9a91c58582a81a1e8197bda WatchSource:0}: Error finding container 03fb71018b6820b54cc186e02849883a08777f4ab9a91c58582a81a1e8197bda: Status 404 returned error can't find the container with id 03fb71018b6820b54cc186e02849883a08777f4ab9a91c58582a81a1e8197bda Dec 03 22:15:51 crc kubenswrapper[4715]: W1203 22:15:51.262662 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60b5688d_8ef1_44f9_b05e_60e99834e994.slice/crio-0950242ed2fa5748695e889bda56e69144f2c62e5c177fc42d255b1c3020f4fd WatchSource:0}: Error finding container 0950242ed2fa5748695e889bda56e69144f2c62e5c177fc42d255b1c3020f4fd: Status 404 returned error can't find the container with id 0950242ed2fa5748695e889bda56e69144f2c62e5c177fc42d255b1c3020f4fd Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.281936 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-hg7kd"] Dec 03 22:15:51 crc kubenswrapper[4715]: W1203 22:15:51.293086 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78c3f6f3_ad41_4173_ad77_b6b2fa76d6b4.slice/crio-85cf53b7ef64f8581df48a0824935187142ddd9c4d9912d22fbc03a8013ba7e8 WatchSource:0}: Error finding container 85cf53b7ef64f8581df48a0824935187142ddd9c4d9912d22fbc03a8013ba7e8: Status 404 returned error can't find the container with id 85cf53b7ef64f8581df48a0824935187142ddd9c4d9912d22fbc03a8013ba7e8 Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.306134 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-d7jfj"] Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.399308 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-e283-account-create-update-r7wc9"] Dec 03 22:15:51 crc kubenswrapper[4715]: W1203 22:15:51.415992 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79758367_7636_4dcd_9026_b4008520c52d.slice/crio-dc289867195aab082a409f9c0a9f35d44ca4ee84e90408d3a03cfc43110e4705 WatchSource:0}: Error finding container dc289867195aab082a409f9c0a9f35d44ca4ee84e90408d3a03cfc43110e4705: Status 404 returned error can't find the container with id dc289867195aab082a409f9c0a9f35d44ca4ee84e90408d3a03cfc43110e4705 Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.869708 4715 generic.go:334] "Generic (PLEG): container finished" podID="075e50c6-4819-42c7-88b3-a73a7f959f81" containerID="d2bd265d54c25da8c6955025d08aa8864ef8690abc4a1968a9b1864bbd709c7d" exitCode=0 Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.869767 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-de87-account-create-update-4tn28" event={"ID":"075e50c6-4819-42c7-88b3-a73a7f959f81","Type":"ContainerDied","Data":"d2bd265d54c25da8c6955025d08aa8864ef8690abc4a1968a9b1864bbd709c7d"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.869838 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-de87-account-create-update-4tn28" event={"ID":"075e50c6-4819-42c7-88b3-a73a7f959f81","Type":"ContainerStarted","Data":"03fb71018b6820b54cc186e02849883a08777f4ab9a91c58582a81a1e8197bda"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.871721 4715 generic.go:334] "Generic (PLEG): container finished" podID="78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4" containerID="68962d8f1e4c9519115b629e15aa9a5c90993e36f0d3dc9b74e72ad203cbb42d" exitCode=0 Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.871764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hg7kd" event={"ID":"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4","Type":"ContainerDied","Data":"68962d8f1e4c9519115b629e15aa9a5c90993e36f0d3dc9b74e72ad203cbb42d"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.871831 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hg7kd" event={"ID":"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4","Type":"ContainerStarted","Data":"85cf53b7ef64f8581df48a0824935187142ddd9c4d9912d22fbc03a8013ba7e8"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.873644 4715 generic.go:334] "Generic (PLEG): container finished" podID="f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7" containerID="535f40006dbf7561ccd0f094557951784b4e3ae0e26192d9b1ab42c3bcc862a2" exitCode=0 Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.873730 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fd69-account-create-update-n7kqs" event={"ID":"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7","Type":"ContainerDied","Data":"535f40006dbf7561ccd0f094557951784b4e3ae0e26192d9b1ab42c3bcc862a2"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.873767 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fd69-account-create-update-n7kqs" event={"ID":"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7","Type":"ContainerStarted","Data":"7c116a2f8f996690099acfe86b930445ded2bdb794dc5514f48b357034d2325e"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.878143 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e283-account-create-update-r7wc9" event={"ID":"79758367-7636-4dcd-9026-b4008520c52d","Type":"ContainerStarted","Data":"0b33a3889d97525c66bbdfebd8e359f643585e31acd484a171e876472aeb31f1"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.878175 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e283-account-create-update-r7wc9" event={"ID":"79758367-7636-4dcd-9026-b4008520c52d","Type":"ContainerStarted","Data":"dc289867195aab082a409f9c0a9f35d44ca4ee84e90408d3a03cfc43110e4705"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.879409 4715 generic.go:334] "Generic (PLEG): container finished" podID="75b6412c-5008-4966-b1d9-55f9846d9cd1" containerID="55cf90793dbe13f14c95414aed7d758928f497a82c05bab05454946788d83cf9" exitCode=0 Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.879554 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j4dtp" event={"ID":"75b6412c-5008-4966-b1d9-55f9846d9cd1","Type":"ContainerDied","Data":"55cf90793dbe13f14c95414aed7d758928f497a82c05bab05454946788d83cf9"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.882133 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d7jfj" event={"ID":"7e7668eb-91d3-4172-9b37-048835ae6899","Type":"ContainerStarted","Data":"a34d862f8e37bf9491374b0a4694a1ebc4b19f21a2390405cd66ddb871c7eb36"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.884044 4715 generic.go:334] "Generic (PLEG): container finished" podID="60b5688d-8ef1-44f9-b05e-60e99834e994" containerID="346469aa378873aea0de480d2af9b2f14c2fd22f9ca74f327dcae8f8fadab8cf" exitCode=0 Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.884100 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tzppt" event={"ID":"60b5688d-8ef1-44f9-b05e-60e99834e994","Type":"ContainerDied","Data":"346469aa378873aea0de480d2af9b2f14c2fd22f9ca74f327dcae8f8fadab8cf"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.884128 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tzppt" event={"ID":"60b5688d-8ef1-44f9-b05e-60e99834e994","Type":"ContainerStarted","Data":"0950242ed2fa5748695e889bda56e69144f2c62e5c177fc42d255b1c3020f4fd"} Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.914239 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-e283-account-create-update-r7wc9" podStartSLOduration=1.914218022 podStartE2EDuration="1.914218022s" podCreationTimestamp="2025-12-03 22:15:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:15:51.907612045 +0000 UTC m=+2048.650322660" watchObservedRunningTime="2025-12-03 22:15:51.914218022 +0000 UTC m=+2048.656928617" Dec 03 22:15:51 crc kubenswrapper[4715]: I1203 22:15:51.924809 4715 scope.go:117] "RemoveContainer" containerID="b8079d9db0601a1f5e4bc6ab874c873fc363425ea45c0846f5af709a2beb3626" Dec 03 22:15:52 crc kubenswrapper[4715]: I1203 22:15:52.896402 4715 generic.go:334] "Generic (PLEG): container finished" podID="79758367-7636-4dcd-9026-b4008520c52d" containerID="0b33a3889d97525c66bbdfebd8e359f643585e31acd484a171e876472aeb31f1" exitCode=0 Dec 03 22:15:52 crc kubenswrapper[4715]: I1203 22:15:52.896492 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e283-account-create-update-r7wc9" event={"ID":"79758367-7636-4dcd-9026-b4008520c52d","Type":"ContainerDied","Data":"0b33a3889d97525c66bbdfebd8e359f643585e31acd484a171e876472aeb31f1"} Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.363637 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.532207 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-operator-scripts\") pod \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\" (UID: \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.532468 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8w8l\" (UniqueName: \"kubernetes.io/projected/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-kube-api-access-v8w8l\") pod \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\" (UID: \"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.532832 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4" (UID: "78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.532937 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.539990 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-kube-api-access-v8w8l" (OuterVolumeSpecName: "kube-api-access-v8w8l") pod "78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4" (UID: "78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4"). InnerVolumeSpecName "kube-api-access-v8w8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.602521 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.612680 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.629469 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.638654 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8w8l\" (UniqueName: \"kubernetes.io/projected/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4-kube-api-access-v8w8l\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.692002 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.739658 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b6412c-5008-4966-b1d9-55f9846d9cd1-operator-scripts\") pod \"75b6412c-5008-4966-b1d9-55f9846d9cd1\" (UID: \"75b6412c-5008-4966-b1d9-55f9846d9cd1\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.739735 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42p9w\" (UniqueName: \"kubernetes.io/projected/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-kube-api-access-42p9w\") pod \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\" (UID: \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.739759 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b5688d-8ef1-44f9-b05e-60e99834e994-operator-scripts\") pod \"60b5688d-8ef1-44f9-b05e-60e99834e994\" (UID: \"60b5688d-8ef1-44f9-b05e-60e99834e994\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.739797 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl78h\" (UniqueName: \"kubernetes.io/projected/60b5688d-8ef1-44f9-b05e-60e99834e994-kube-api-access-dl78h\") pod \"60b5688d-8ef1-44f9-b05e-60e99834e994\" (UID: \"60b5688d-8ef1-44f9-b05e-60e99834e994\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.739824 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-operator-scripts\") pod \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\" (UID: \"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.739908 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svbtq\" (UniqueName: \"kubernetes.io/projected/75b6412c-5008-4966-b1d9-55f9846d9cd1-kube-api-access-svbtq\") pod \"75b6412c-5008-4966-b1d9-55f9846d9cd1\" (UID: \"75b6412c-5008-4966-b1d9-55f9846d9cd1\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.740648 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60b5688d-8ef1-44f9-b05e-60e99834e994-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60b5688d-8ef1-44f9-b05e-60e99834e994" (UID: "60b5688d-8ef1-44f9-b05e-60e99834e994"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.740781 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7" (UID: "f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.741480 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75b6412c-5008-4966-b1d9-55f9846d9cd1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "75b6412c-5008-4966-b1d9-55f9846d9cd1" (UID: "75b6412c-5008-4966-b1d9-55f9846d9cd1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.744785 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75b6412c-5008-4966-b1d9-55f9846d9cd1-kube-api-access-svbtq" (OuterVolumeSpecName: "kube-api-access-svbtq") pod "75b6412c-5008-4966-b1d9-55f9846d9cd1" (UID: "75b6412c-5008-4966-b1d9-55f9846d9cd1"). InnerVolumeSpecName "kube-api-access-svbtq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.744933 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60b5688d-8ef1-44f9-b05e-60e99834e994-kube-api-access-dl78h" (OuterVolumeSpecName: "kube-api-access-dl78h") pod "60b5688d-8ef1-44f9-b05e-60e99834e994" (UID: "60b5688d-8ef1-44f9-b05e-60e99834e994"). InnerVolumeSpecName "kube-api-access-dl78h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.744984 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-kube-api-access-42p9w" (OuterVolumeSpecName: "kube-api-access-42p9w") pod "f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7" (UID: "f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7"). InnerVolumeSpecName "kube-api-access-42p9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.841308 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p6h9\" (UniqueName: \"kubernetes.io/projected/075e50c6-4819-42c7-88b3-a73a7f959f81-kube-api-access-8p6h9\") pod \"075e50c6-4819-42c7-88b3-a73a7f959f81\" (UID: \"075e50c6-4819-42c7-88b3-a73a7f959f81\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.841387 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075e50c6-4819-42c7-88b3-a73a7f959f81-operator-scripts\") pod \"075e50c6-4819-42c7-88b3-a73a7f959f81\" (UID: \"075e50c6-4819-42c7-88b3-a73a7f959f81\") " Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.842475 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75b6412c-5008-4966-b1d9-55f9846d9cd1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.842495 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42p9w\" (UniqueName: \"kubernetes.io/projected/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-kube-api-access-42p9w\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.842519 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b5688d-8ef1-44f9-b05e-60e99834e994-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.842530 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl78h\" (UniqueName: \"kubernetes.io/projected/60b5688d-8ef1-44f9-b05e-60e99834e994-kube-api-access-dl78h\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.842541 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.842550 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svbtq\" (UniqueName: \"kubernetes.io/projected/75b6412c-5008-4966-b1d9-55f9846d9cd1-kube-api-access-svbtq\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.842740 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/075e50c6-4819-42c7-88b3-a73a7f959f81-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "075e50c6-4819-42c7-88b3-a73a7f959f81" (UID: "075e50c6-4819-42c7-88b3-a73a7f959f81"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.848673 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/075e50c6-4819-42c7-88b3-a73a7f959f81-kube-api-access-8p6h9" (OuterVolumeSpecName: "kube-api-access-8p6h9") pod "075e50c6-4819-42c7-88b3-a73a7f959f81" (UID: "075e50c6-4819-42c7-88b3-a73a7f959f81"). InnerVolumeSpecName "kube-api-access-8p6h9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.912029 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fd69-account-create-update-n7kqs" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.913536 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fd69-account-create-update-n7kqs" event={"ID":"f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7","Type":"ContainerDied","Data":"7c116a2f8f996690099acfe86b930445ded2bdb794dc5514f48b357034d2325e"} Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.913570 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c116a2f8f996690099acfe86b930445ded2bdb794dc5514f48b357034d2325e" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.915902 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-j4dtp" event={"ID":"75b6412c-5008-4966-b1d9-55f9846d9cd1","Type":"ContainerDied","Data":"758aeb89be7cb6740d2a74aca847aa94d7755e164be384abe39872fef05977ad"} Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.915943 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="758aeb89be7cb6740d2a74aca847aa94d7755e164be384abe39872fef05977ad" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.916000 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-j4dtp" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.919123 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-tzppt" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.919208 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-tzppt" event={"ID":"60b5688d-8ef1-44f9-b05e-60e99834e994","Type":"ContainerDied","Data":"0950242ed2fa5748695e889bda56e69144f2c62e5c177fc42d255b1c3020f4fd"} Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.919281 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0950242ed2fa5748695e889bda56e69144f2c62e5c177fc42d255b1c3020f4fd" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.920702 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-de87-account-create-update-4tn28" event={"ID":"075e50c6-4819-42c7-88b3-a73a7f959f81","Type":"ContainerDied","Data":"03fb71018b6820b54cc186e02849883a08777f4ab9a91c58582a81a1e8197bda"} Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.920728 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-de87-account-create-update-4tn28" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.920737 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03fb71018b6820b54cc186e02849883a08777f4ab9a91c58582a81a1e8197bda" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.922166 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-hg7kd" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.922773 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-hg7kd" event={"ID":"78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4","Type":"ContainerDied","Data":"85cf53b7ef64f8581df48a0824935187142ddd9c4d9912d22fbc03a8013ba7e8"} Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.922802 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85cf53b7ef64f8581df48a0824935187142ddd9c4d9912d22fbc03a8013ba7e8" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.944599 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p6h9\" (UniqueName: \"kubernetes.io/projected/075e50c6-4819-42c7-88b3-a73a7f959f81-kube-api-access-8p6h9\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:53 crc kubenswrapper[4715]: I1203 22:15:53.944640 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075e50c6-4819-42c7-88b3-a73a7f959f81-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:15:56 crc kubenswrapper[4715]: I1203 22:15:56.668799 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:15:56 crc kubenswrapper[4715]: I1203 22:15:56.733389 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bqnl2"] Dec 03 22:15:56 crc kubenswrapper[4715]: I1203 22:15:56.734054 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" podUID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerName="dnsmasq-dns" containerID="cri-o://52e2c8febb319080e0af0f7ffc895517cb374371191d68ea7055ba4ea833dfb5" gracePeriod=10 Dec 03 22:15:57 crc kubenswrapper[4715]: I1203 22:15:57.958366 4715 generic.go:334] "Generic (PLEG): container finished" podID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerID="52e2c8febb319080e0af0f7ffc895517cb374371191d68ea7055ba4ea833dfb5" exitCode=0 Dec 03 22:15:57 crc kubenswrapper[4715]: I1203 22:15:57.958400 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" event={"ID":"793fb19c-ce52-40e2-b874-d0c42cf1ad50","Type":"ContainerDied","Data":"52e2c8febb319080e0af0f7ffc895517cb374371191d68ea7055ba4ea833dfb5"} Dec 03 22:16:00 crc kubenswrapper[4715]: I1203 22:16:00.462275 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" podUID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.115:5353: connect: connection refused" Dec 03 22:16:02 crc kubenswrapper[4715]: I1203 22:16:02.785434 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:16:02 crc kubenswrapper[4715]: I1203 22:16:02.802683 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79758367-7636-4dcd-9026-b4008520c52d-operator-scripts\") pod \"79758367-7636-4dcd-9026-b4008520c52d\" (UID: \"79758367-7636-4dcd-9026-b4008520c52d\") " Dec 03 22:16:02 crc kubenswrapper[4715]: I1203 22:16:02.803079 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rvjb\" (UniqueName: \"kubernetes.io/projected/79758367-7636-4dcd-9026-b4008520c52d-kube-api-access-5rvjb\") pod \"79758367-7636-4dcd-9026-b4008520c52d\" (UID: \"79758367-7636-4dcd-9026-b4008520c52d\") " Dec 03 22:16:02 crc kubenswrapper[4715]: I1203 22:16:02.803391 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79758367-7636-4dcd-9026-b4008520c52d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79758367-7636-4dcd-9026-b4008520c52d" (UID: "79758367-7636-4dcd-9026-b4008520c52d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:02 crc kubenswrapper[4715]: I1203 22:16:02.810875 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79758367-7636-4dcd-9026-b4008520c52d-kube-api-access-5rvjb" (OuterVolumeSpecName: "kube-api-access-5rvjb") pod "79758367-7636-4dcd-9026-b4008520c52d" (UID: "79758367-7636-4dcd-9026-b4008520c52d"). InnerVolumeSpecName "kube-api-access-5rvjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:02 crc kubenswrapper[4715]: I1203 22:16:02.904133 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79758367-7636-4dcd-9026-b4008520c52d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:02 crc kubenswrapper[4715]: I1203 22:16:02.904180 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rvjb\" (UniqueName: \"kubernetes.io/projected/79758367-7636-4dcd-9026-b4008520c52d-kube-api-access-5rvjb\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:02 crc kubenswrapper[4715]: I1203 22:16:02.958839 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.005059 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-config\") pod \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.005202 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8zlt\" (UniqueName: \"kubernetes.io/projected/793fb19c-ce52-40e2-b874-d0c42cf1ad50-kube-api-access-n8zlt\") pod \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.005246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-dns-svc\") pod \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.005287 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-sb\") pod \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.005315 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-nb\") pod \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\" (UID: \"793fb19c-ce52-40e2-b874-d0c42cf1ad50\") " Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.010157 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/793fb19c-ce52-40e2-b874-d0c42cf1ad50-kube-api-access-n8zlt" (OuterVolumeSpecName: "kube-api-access-n8zlt") pod "793fb19c-ce52-40e2-b874-d0c42cf1ad50" (UID: "793fb19c-ce52-40e2-b874-d0c42cf1ad50"). InnerVolumeSpecName "kube-api-access-n8zlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.027995 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.027998 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-bqnl2" event={"ID":"793fb19c-ce52-40e2-b874-d0c42cf1ad50","Type":"ContainerDied","Data":"076ffb1543443d2cb189099533b094e6217f904304eed93f0afec356105ba44c"} Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.028112 4715 scope.go:117] "RemoveContainer" containerID="52e2c8febb319080e0af0f7ffc895517cb374371191d68ea7055ba4ea833dfb5" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.030175 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-e283-account-create-update-r7wc9" event={"ID":"79758367-7636-4dcd-9026-b4008520c52d","Type":"ContainerDied","Data":"dc289867195aab082a409f9c0a9f35d44ca4ee84e90408d3a03cfc43110e4705"} Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.030201 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc289867195aab082a409f9c0a9f35d44ca4ee84e90408d3a03cfc43110e4705" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.030226 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-e283-account-create-update-r7wc9" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.031978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d7jfj" event={"ID":"7e7668eb-91d3-4172-9b37-048835ae6899","Type":"ContainerStarted","Data":"54ea093b6c7ea48d7d2d786ef880b3512293ff422760c010dd224c8392081bae"} Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.047287 4715 scope.go:117] "RemoveContainer" containerID="e9c9503666ff888202257ffb5e64214f9ec6908e5c1680cb6d1ea99842a92cf2" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.055881 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "793fb19c-ce52-40e2-b874-d0c42cf1ad50" (UID: "793fb19c-ce52-40e2-b874-d0c42cf1ad50"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.067940 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "793fb19c-ce52-40e2-b874-d0c42cf1ad50" (UID: "793fb19c-ce52-40e2-b874-d0c42cf1ad50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.069362 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-d7jfj" podStartSLOduration=1.680889852 podStartE2EDuration="13.069346742s" podCreationTimestamp="2025-12-03 22:15:50 +0000 UTC" firstStartedPulling="2025-12-03 22:15:51.331641623 +0000 UTC m=+2048.074352218" lastFinishedPulling="2025-12-03 22:16:02.720098483 +0000 UTC m=+2059.462809108" observedRunningTime="2025-12-03 22:16:03.045942127 +0000 UTC m=+2059.788652732" watchObservedRunningTime="2025-12-03 22:16:03.069346742 +0000 UTC m=+2059.812057337" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.076640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "793fb19c-ce52-40e2-b874-d0c42cf1ad50" (UID: "793fb19c-ce52-40e2-b874-d0c42cf1ad50"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.078734 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-config" (OuterVolumeSpecName: "config") pod "793fb19c-ce52-40e2-b874-d0c42cf1ad50" (UID: "793fb19c-ce52-40e2-b874-d0c42cf1ad50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.107228 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.107261 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.107276 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.107288 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8zlt\" (UniqueName: \"kubernetes.io/projected/793fb19c-ce52-40e2-b874-d0c42cf1ad50-kube-api-access-n8zlt\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.107303 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/793fb19c-ce52-40e2-b874-d0c42cf1ad50-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.382488 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bqnl2"] Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.396529 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-bqnl2"] Dec 03 22:16:03 crc kubenswrapper[4715]: I1203 22:16:03.653245 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" path="/var/lib/kubelet/pods/793fb19c-ce52-40e2-b874-d0c42cf1ad50/volumes" Dec 03 22:16:05 crc kubenswrapper[4715]: I1203 22:16:05.160430 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:16:05 crc kubenswrapper[4715]: I1203 22:16:05.160597 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:16:05 crc kubenswrapper[4715]: I1203 22:16:05.160682 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:16:05 crc kubenswrapper[4715]: I1203 22:16:05.161651 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a7b507f93c05d0aab4e197e7602419073e36d87d0fa5a60853ef24a637d4567"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:16:05 crc kubenswrapper[4715]: I1203 22:16:05.161776 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://2a7b507f93c05d0aab4e197e7602419073e36d87d0fa5a60853ef24a637d4567" gracePeriod=600 Dec 03 22:16:06 crc kubenswrapper[4715]: I1203 22:16:06.062411 4715 generic.go:334] "Generic (PLEG): container finished" podID="7e7668eb-91d3-4172-9b37-048835ae6899" containerID="54ea093b6c7ea48d7d2d786ef880b3512293ff422760c010dd224c8392081bae" exitCode=0 Dec 03 22:16:06 crc kubenswrapper[4715]: I1203 22:16:06.062512 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d7jfj" event={"ID":"7e7668eb-91d3-4172-9b37-048835ae6899","Type":"ContainerDied","Data":"54ea093b6c7ea48d7d2d786ef880b3512293ff422760c010dd224c8392081bae"} Dec 03 22:16:06 crc kubenswrapper[4715]: I1203 22:16:06.065650 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="2a7b507f93c05d0aab4e197e7602419073e36d87d0fa5a60853ef24a637d4567" exitCode=0 Dec 03 22:16:06 crc kubenswrapper[4715]: I1203 22:16:06.065682 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"2a7b507f93c05d0aab4e197e7602419073e36d87d0fa5a60853ef24a637d4567"} Dec 03 22:16:06 crc kubenswrapper[4715]: I1203 22:16:06.065704 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc"} Dec 03 22:16:06 crc kubenswrapper[4715]: I1203 22:16:06.065720 4715 scope.go:117] "RemoveContainer" containerID="d4e8d342a0569c7f635917479d8a0e1c104c42eea3e37cf5bf618b64d96b4e02" Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.513783 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.695449 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-combined-ca-bundle\") pod \"7e7668eb-91d3-4172-9b37-048835ae6899\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.695684 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wsc7\" (UniqueName: \"kubernetes.io/projected/7e7668eb-91d3-4172-9b37-048835ae6899-kube-api-access-5wsc7\") pod \"7e7668eb-91d3-4172-9b37-048835ae6899\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.695846 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-config-data\") pod \"7e7668eb-91d3-4172-9b37-048835ae6899\" (UID: \"7e7668eb-91d3-4172-9b37-048835ae6899\") " Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.706342 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e7668eb-91d3-4172-9b37-048835ae6899-kube-api-access-5wsc7" (OuterVolumeSpecName: "kube-api-access-5wsc7") pod "7e7668eb-91d3-4172-9b37-048835ae6899" (UID: "7e7668eb-91d3-4172-9b37-048835ae6899"). InnerVolumeSpecName "kube-api-access-5wsc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.745200 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e7668eb-91d3-4172-9b37-048835ae6899" (UID: "7e7668eb-91d3-4172-9b37-048835ae6899"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.761409 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-config-data" (OuterVolumeSpecName: "config-data") pod "7e7668eb-91d3-4172-9b37-048835ae6899" (UID: "7e7668eb-91d3-4172-9b37-048835ae6899"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.798481 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.798550 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e7668eb-91d3-4172-9b37-048835ae6899-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:07 crc kubenswrapper[4715]: I1203 22:16:07.798572 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wsc7\" (UniqueName: \"kubernetes.io/projected/7e7668eb-91d3-4172-9b37-048835ae6899-kube-api-access-5wsc7\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.094657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-d7jfj" event={"ID":"7e7668eb-91d3-4172-9b37-048835ae6899","Type":"ContainerDied","Data":"a34d862f8e37bf9491374b0a4694a1ebc4b19f21a2390405cd66ddb871c7eb36"} Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.095169 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a34d862f8e37bf9491374b0a4694a1ebc4b19f21a2390405cd66ddb871c7eb36" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.094850 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-d7jfj" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.398846 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-pgkfz"] Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399238 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e7668eb-91d3-4172-9b37-048835ae6899" containerName="keystone-db-sync" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399263 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e7668eb-91d3-4172-9b37-048835ae6899" containerName="keystone-db-sync" Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399280 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerName="init" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399286 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerName="init" Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399295 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="075e50c6-4819-42c7-88b3-a73a7f959f81" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399302 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="075e50c6-4819-42c7-88b3-a73a7f959f81" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399310 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399315 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399323 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399329 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399340 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75b6412c-5008-4966-b1d9-55f9846d9cd1" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399347 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="75b6412c-5008-4966-b1d9-55f9846d9cd1" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399358 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b5688d-8ef1-44f9-b05e-60e99834e994" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399364 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b5688d-8ef1-44f9-b05e-60e99834e994" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399376 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79758367-7636-4dcd-9026-b4008520c52d" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399382 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="79758367-7636-4dcd-9026-b4008520c52d" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: E1203 22:16:08.399392 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerName="dnsmasq-dns" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399398 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerName="dnsmasq-dns" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399561 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="793fb19c-ce52-40e2-b874-d0c42cf1ad50" containerName="dnsmasq-dns" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399576 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="75b6412c-5008-4966-b1d9-55f9846d9cd1" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399586 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="075e50c6-4819-42c7-88b3-a73a7f959f81" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399596 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b5688d-8ef1-44f9-b05e-60e99834e994" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399605 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4" containerName="mariadb-database-create" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399617 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="79758367-7636-4dcd-9026-b4008520c52d" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399626 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e7668eb-91d3-4172-9b37-048835ae6899" containerName="keystone-db-sync" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.399634 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7" containerName="mariadb-account-create-update" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.406448 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.413384 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9bcdn" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.413580 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.413389 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.413756 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.418051 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.425296 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-x92wb"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.426805 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.451060 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pgkfz"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.499575 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-x92wb"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511433 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-config\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511482 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-scripts\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511517 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks8gn\" (UniqueName: \"kubernetes.io/projected/5a7bfab0-bbd0-4773-a92a-54449b1f486b-kube-api-access-ks8gn\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511541 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-fernet-keys\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511561 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511576 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511597 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-credential-keys\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511614 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ltjt\" (UniqueName: \"kubernetes.io/projected/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-kube-api-access-8ltjt\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511635 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-combined-ca-bundle\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511671 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511693 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.511723 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-config-data\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.571492 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-blfqt"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.577660 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.584544 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-n9b5g" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.588913 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.588978 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.591637 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-blfqt"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.614532 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.614790 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.614880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-config-data\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.614979 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-config\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615086 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-scripts\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615161 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks8gn\" (UniqueName: \"kubernetes.io/projected/5a7bfab0-bbd0-4773-a92a-54449b1f486b-kube-api-access-ks8gn\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615234 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-fernet-keys\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615307 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615372 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615458 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-credential-keys\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615548 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ltjt\" (UniqueName: \"kubernetes.io/projected/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-kube-api-access-8ltjt\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615622 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-combined-ca-bundle\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615623 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.616264 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-config\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.617172 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.617735 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.615557 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.627496 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-combined-ca-bundle\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.628317 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-fernet-keys\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.633927 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-scripts\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.648849 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-config-data\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.657382 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-credential-keys\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.657452 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-55967db74f-69gh7"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.658992 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.659077 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks8gn\" (UniqueName: \"kubernetes.io/projected/5a7bfab0-bbd0-4773-a92a-54449b1f486b-kube-api-access-ks8gn\") pod \"keystone-bootstrap-pgkfz\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.667554 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55967db74f-69gh7"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.679453 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.679522 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.679535 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ltjt\" (UniqueName: \"kubernetes.io/projected/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-kube-api-access-8ltjt\") pod \"dnsmasq-dns-bbf5cc879-x92wb\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.679674 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-gm2dm" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.679682 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.720442 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-combined-ca-bundle\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.720782 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8kmz\" (UniqueName: \"kubernetes.io/projected/6ded23d0-fc8b-4ed7-8358-573da1f24919-kube-api-access-c8kmz\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.720865 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-config\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.724843 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.726867 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.735772 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.736012 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.736851 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.766895 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.772591 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-d8zcz"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.774041 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.785019 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rhz28" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.785267 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.785527 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.792744 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.820562 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-6ngmv"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.821674 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.823317 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-config\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.823377 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-horizon-secret-key\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.823407 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-scripts\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.823458 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-combined-ca-bundle\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827340 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-log-httpd\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827390 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57rms\" (UniqueName: \"kubernetes.io/projected/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-kube-api-access-57rms\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827407 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827426 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827443 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-config-data\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827520 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-config-data\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827543 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-logs\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827618 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-run-httpd\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827683 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-scripts\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827723 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8kmz\" (UniqueName: \"kubernetes.io/projected/6ded23d0-fc8b-4ed7-8358-573da1f24919-kube-api-access-c8kmz\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.827739 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lp2b\" (UniqueName: \"kubernetes.io/projected/72b03167-1df1-47c3-94c7-e0f13b65f972-kube-api-access-2lp2b\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.858046 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-combined-ca-bundle\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.858615 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.858955 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mwckq" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.863397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-config\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.863470 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-d8zcz"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.877627 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8kmz\" (UniqueName: \"kubernetes.io/projected/6ded23d0-fc8b-4ed7-8358-573da1f24919-kube-api-access-c8kmz\") pod \"neutron-db-sync-blfqt\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.885575 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6ngmv"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.897386 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.910359 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-x92wb"] Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930170 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-config-data\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930260 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-log-httpd\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930288 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-scripts\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57rms\" (UniqueName: \"kubernetes.io/projected/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-kube-api-access-57rms\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930341 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930362 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930382 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-config-data\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-combined-ca-bundle\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930441 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-config-data\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930463 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-logs\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930724 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-combined-ca-bundle\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930777 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsskp\" (UniqueName: \"kubernetes.io/projected/2eba90e4-4991-4260-9e8e-2a3a5412b264-kube-api-access-dsskp\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930803 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e2fa00-220e-4811-8b41-644d96b70a46-logs\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930835 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgtrr\" (UniqueName: \"kubernetes.io/projected/d3e2fa00-220e-4811-8b41-644d96b70a46-kube-api-access-cgtrr\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930857 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-run-httpd\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930903 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-scripts\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930931 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lp2b\" (UniqueName: \"kubernetes.io/projected/72b03167-1df1-47c3-94c7-e0f13b65f972-kube-api-access-2lp2b\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930967 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-db-sync-config-data\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.930991 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-horizon-secret-key\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.931026 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-scripts\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.934930 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-log-httpd\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.935603 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-scripts\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.935905 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-run-httpd\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.942286 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-logs\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.944327 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-scripts\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.946288 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-horizon-secret-key\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.948818 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-config-data\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.973212 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lp2b\" (UniqueName: \"kubernetes.io/projected/72b03167-1df1-47c3-94c7-e0f13b65f972-kube-api-access-2lp2b\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.973585 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.974124 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57rms\" (UniqueName: \"kubernetes.io/projected/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-kube-api-access-57rms\") pod \"horizon-55967db74f-69gh7\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:08 crc kubenswrapper[4715]: I1203 22:16:08.982107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.008373 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-config-data\") pod \"ceilometer-0\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " pod="openstack/ceilometer-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.032588 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-p4mjk"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.040298 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.039921 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-scripts\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.041981 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-combined-ca-bundle\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.042070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-combined-ca-bundle\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.042164 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsskp\" (UniqueName: \"kubernetes.io/projected/2eba90e4-4991-4260-9e8e-2a3a5412b264-kube-api-access-dsskp\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.042229 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e2fa00-220e-4811-8b41-644d96b70a46-logs\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.042288 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgtrr\" (UniqueName: \"kubernetes.io/projected/d3e2fa00-220e-4811-8b41-644d96b70a46-kube-api-access-cgtrr\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.042413 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-db-sync-config-data\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.042895 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-config-data\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.046015 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.046323 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.046590 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wb6bd" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.049800 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-scripts\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.050644 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e2fa00-220e-4811-8b41-644d96b70a46-logs\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.051554 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-combined-ca-bundle\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.057736 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-p4mjk"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.064387 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-db-sync-config-data\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.071629 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.072636 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgtrr\" (UniqueName: \"kubernetes.io/projected/d3e2fa00-220e-4811-8b41-644d96b70a46-kube-api-access-cgtrr\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.074790 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-config-data\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.075834 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-combined-ca-bundle\") pod \"placement-db-sync-d8zcz\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.077244 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsskp\" (UniqueName: \"kubernetes.io/projected/2eba90e4-4991-4260-9e8e-2a3a5412b264-kube-api-access-dsskp\") pod \"barbican-db-sync-6ngmv\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.082534 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zf69s"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.089091 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.094014 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zf69s"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.102453 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.103950 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.105271 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.106260 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.112530 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.112952 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qcrqd" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.119972 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-868658b5-6jkj6"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.121415 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.126726 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.137603 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-868658b5-6jkj6"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.148888 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.150509 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.154074 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.154669 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.157372 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.158345 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.160444 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec32c20-9117-4eca-9b73-a7f8b84677da-etc-machine-id\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.160615 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9t27x\" (UniqueName: \"kubernetes.io/projected/5ec32c20-9117-4eca-9b73-a7f8b84677da-kube-api-access-9t27x\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.160716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-combined-ca-bundle\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.160796 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-db-sync-config-data\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.160868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-config-data\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.160963 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-scripts\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.212146 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.231832 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262660 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-config-data\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262737 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-combined-ca-bundle\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262787 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grj5v\" (UniqueName: \"kubernetes.io/projected/e05e03dd-8c4f-423c-bfba-db6d10fef841-kube-api-access-grj5v\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262811 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-db-sync-config-data\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262832 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-config-data\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262861 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dfb98490-b33a-46a6-9f7d-668862bd0430-horizon-secret-key\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262914 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-config-data\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262933 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-logs\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262966 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-config\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.262989 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vklm5\" (UniqueName: \"kubernetes.io/projected/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-kube-api-access-vklm5\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-scripts\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263050 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263083 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb98490-b33a-46a6-9f7d-668862bd0430-logs\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263106 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263135 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-scripts\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263161 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-logs\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263185 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec32c20-9117-4eca-9b73-a7f8b84677da-etc-machine-id\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263230 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263259 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263284 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263324 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263364 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263394 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263427 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263461 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxfvz\" (UniqueName: \"kubernetes.io/projected/391549ec-5cda-4f77-917c-0b485e1d57bf-kube-api-access-gxfvz\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263494 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-scripts\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263590 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4ch7\" (UniqueName: \"kubernetes.io/projected/dfb98490-b33a-46a6-9f7d-668862bd0430-kube-api-access-n4ch7\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263617 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263650 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9t27x\" (UniqueName: \"kubernetes.io/projected/5ec32c20-9117-4eca-9b73-a7f8b84677da-kube-api-access-9t27x\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263684 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263713 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.263892 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec32c20-9117-4eca-9b73-a7f8b84677da-etc-machine-id\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.268382 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-db-sync-config-data\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.268812 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-config-data\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.269319 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-combined-ca-bundle\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.271978 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-scripts\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.284171 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9t27x\" (UniqueName: \"kubernetes.io/projected/5ec32c20-9117-4eca-9b73-a7f8b84677da-kube-api-access-9t27x\") pod \"cinder-db-sync-p4mjk\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366294 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366371 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366402 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb98490-b33a-46a6-9f7d-668862bd0430-logs\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366435 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-scripts\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366476 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-logs\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366499 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366552 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366605 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366660 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366693 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366729 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366787 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxfvz\" (UniqueName: \"kubernetes.io/projected/391549ec-5cda-4f77-917c-0b485e1d57bf-kube-api-access-gxfvz\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366823 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-scripts\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366867 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366898 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4ch7\" (UniqueName: \"kubernetes.io/projected/dfb98490-b33a-46a6-9f7d-668862bd0430-kube-api-access-n4ch7\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366924 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366965 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.366997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.367022 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-config-data\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.367050 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grj5v\" (UniqueName: \"kubernetes.io/projected/e05e03dd-8c4f-423c-bfba-db6d10fef841-kube-api-access-grj5v\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.367081 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dfb98490-b33a-46a6-9f7d-668862bd0430-horizon-secret-key\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.367113 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-config-data\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.367141 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-logs\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.367173 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-config\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.367197 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vklm5\" (UniqueName: \"kubernetes.io/projected/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-kube-api-access-vklm5\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.371464 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.371660 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.372388 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.372685 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-logs\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.373208 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.373353 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-config\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.373558 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-config-data\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.374135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-logs\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.374385 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb98490-b33a-46a6-9f7d-668862bd0430-logs\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.374639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.374838 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-scripts\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.375607 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.375706 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.375723 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.379722 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-config-data\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.382291 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.394123 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vklm5\" (UniqueName: \"kubernetes.io/projected/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-kube-api-access-vklm5\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.394486 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.408189 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-scripts\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.409652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dfb98490-b33a-46a6-9f7d-668862bd0430-horizon-secret-key\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.411428 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-config-data\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.415358 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.416977 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-scripts\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.418130 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxfvz\" (UniqueName: \"kubernetes.io/projected/391549ec-5cda-4f77-917c-0b485e1d57bf-kube-api-access-gxfvz\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.418343 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.419846 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grj5v\" (UniqueName: \"kubernetes.io/projected/e05e03dd-8c4f-423c-bfba-db6d10fef841-kube-api-access-grj5v\") pod \"dnsmasq-dns-56df8fb6b7-zf69s\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.423691 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-pgkfz"] Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.424218 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4ch7\" (UniqueName: \"kubernetes.io/projected/dfb98490-b33a-46a6-9f7d-668862bd0430-kube-api-access-n4ch7\") pod \"horizon-868658b5-6jkj6\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.426724 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.437928 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.439893 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.455070 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.496812 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.510255 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.513711 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:09 crc kubenswrapper[4715]: I1203 22:16:09.688352 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-x92wb"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.002723 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55967db74f-69gh7"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.037750 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:16:10 crc kubenswrapper[4715]: W1203 22:16:10.083018 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98853dba_6c7a_4ed5_afa4_3696e1e1eee6.slice/crio-84225bacf08c0f9534733f5ea0c81fb190f90ee005f8b6c9ccd62ff18b9a831c WatchSource:0}: Error finding container 84225bacf08c0f9534733f5ea0c81fb190f90ee005f8b6c9ccd62ff18b9a831c: Status 404 returned error can't find the container with id 84225bacf08c0f9534733f5ea0c81fb190f90ee005f8b6c9ccd62ff18b9a831c Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.113853 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-blfqt"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.148120 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55967db74f-69gh7" event={"ID":"98853dba-6c7a-4ed5-afa4-3696e1e1eee6","Type":"ContainerStarted","Data":"84225bacf08c0f9534733f5ea0c81fb190f90ee005f8b6c9ccd62ff18b9a831c"} Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.149722 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-blfqt" event={"ID":"6ded23d0-fc8b-4ed7-8358-573da1f24919","Type":"ContainerStarted","Data":"de781bce412099648733b54ebfd61580f0ee94df7e9a0af6e62cdaaf1b7d4a11"} Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.151154 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72b03167-1df1-47c3-94c7-e0f13b65f972","Type":"ContainerStarted","Data":"8e1527e2a13e8a9a7188b887d9741cfc81f3bb52816d223d553d52ba2fa0d068"} Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.156341 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pgkfz" event={"ID":"5a7bfab0-bbd0-4773-a92a-54449b1f486b","Type":"ContainerStarted","Data":"c4a39a7dad139e08366ecbf387d73e370b5ef182fc0505f732aeb4cdce19a27a"} Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.161748 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" event={"ID":"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869","Type":"ContainerStarted","Data":"4425df45bc1a3c941958509623fa74a8437396092cb60fc60a6494280bcf99e4"} Dec 03 22:16:10 crc kubenswrapper[4715]: W1203 22:16:10.230589 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3e2fa00_220e_4811_8b41_644d96b70a46.slice/crio-1e1b216fe46d6b2ff507fda8255f6644a7f51d023f7bfd4307c0feb95ba11f59 WatchSource:0}: Error finding container 1e1b216fe46d6b2ff507fda8255f6644a7f51d023f7bfd4307c0feb95ba11f59: Status 404 returned error can't find the container with id 1e1b216fe46d6b2ff507fda8255f6644a7f51d023f7bfd4307c0feb95ba11f59 Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.243990 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-d8zcz"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.251199 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-6ngmv"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.260734 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-p4mjk"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.406717 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zf69s"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.476428 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.507295 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-868658b5-6jkj6"] Dec 03 22:16:10 crc kubenswrapper[4715]: W1203 22:16:10.523254 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfb98490_b33a_46a6_9f7d_668862bd0430.slice/crio-c9427c5d66d898d1c1c37665df0f0a69ba873b7702ae617452dc51c42ed83ebe WatchSource:0}: Error finding container c9427c5d66d898d1c1c37665df0f0a69ba873b7702ae617452dc51c42ed83ebe: Status 404 returned error can't find the container with id c9427c5d66d898d1c1c37665df0f0a69ba873b7702ae617452dc51c42ed83ebe Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.647394 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.763423 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.773904 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55967db74f-69gh7"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.807468 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6d97bc8c5c-g2ds9"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.809184 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.844130 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.875231 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d97bc8c5c-g2ds9"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.919817 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.947771 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-logs\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.947863 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-config-data\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.947891 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-horizon-secret-key\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.947933 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-scripts\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:10 crc kubenswrapper[4715]: I1203 22:16:10.947982 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dxjs\" (UniqueName: \"kubernetes.io/projected/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-kube-api-access-9dxjs\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.053898 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dxjs\" (UniqueName: \"kubernetes.io/projected/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-kube-api-access-9dxjs\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.054016 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-logs\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.054070 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-config-data\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.054086 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-horizon-secret-key\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.054118 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-scripts\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.054873 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-logs\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.055149 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-scripts\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.055647 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-config-data\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.059458 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-horizon-secret-key\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.070718 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dxjs\" (UniqueName: \"kubernetes.io/projected/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-kube-api-access-9dxjs\") pod \"horizon-6d97bc8c5c-g2ds9\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.147249 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.186450 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec","Type":"ContainerStarted","Data":"6e90802a0e9412beed9c3f0b46f3b07af8b09aa86b0f3961e061070013f4b613"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.187345 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"391549ec-5cda-4f77-917c-0b485e1d57bf","Type":"ContainerStarted","Data":"8858e04e32485c764a9ff25c65cfc910575f2df7dac66cff0e4171abd71c0fe8"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.189466 4715 generic.go:334] "Generic (PLEG): container finished" podID="b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" containerID="24720ddd54bbf1945f48c8116231a134e5ca1e8776a928c63eb38ceecac14c8c" exitCode=0 Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.189625 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" event={"ID":"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869","Type":"ContainerDied","Data":"24720ddd54bbf1945f48c8116231a134e5ca1e8776a928c63eb38ceecac14c8c"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.193843 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6ngmv" event={"ID":"2eba90e4-4991-4260-9e8e-2a3a5412b264","Type":"ContainerStarted","Data":"4e7d91578a5e8d5d9ed60a8225035599b07e3a00d2b8f3477a7cf87554210224"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.195435 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-868658b5-6jkj6" event={"ID":"dfb98490-b33a-46a6-9f7d-668862bd0430","Type":"ContainerStarted","Data":"c9427c5d66d898d1c1c37665df0f0a69ba873b7702ae617452dc51c42ed83ebe"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.222450 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-blfqt" event={"ID":"6ded23d0-fc8b-4ed7-8358-573da1f24919","Type":"ContainerStarted","Data":"371aa80cbef5669afbd1d3dc6067b637a7c1618fca9e34f0e35de061c608b7c8"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.235118 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pgkfz" event={"ID":"5a7bfab0-bbd0-4773-a92a-54449b1f486b","Type":"ContainerStarted","Data":"a5b0d903b1f711b99655d67b5066c14cddb95948e7c959d7124c9576077583ba"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.235910 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-blfqt" podStartSLOduration=3.235892507 podStartE2EDuration="3.235892507s" podCreationTimestamp="2025-12-03 22:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:16:11.234924341 +0000 UTC m=+2067.977634936" watchObservedRunningTime="2025-12-03 22:16:11.235892507 +0000 UTC m=+2067.978603102" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.255773 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-pgkfz" podStartSLOduration=3.255754398 podStartE2EDuration="3.255754398s" podCreationTimestamp="2025-12-03 22:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:16:11.250724194 +0000 UTC m=+2067.993434789" watchObservedRunningTime="2025-12-03 22:16:11.255754398 +0000 UTC m=+2067.998464993" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.276369 4715 generic.go:334] "Generic (PLEG): container finished" podID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerID="71de947a8dec65c3ee92f5feadf901bb7d74f27f725fdfb0d808ce3a07822f50" exitCode=0 Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.276487 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" event={"ID":"e05e03dd-8c4f-423c-bfba-db6d10fef841","Type":"ContainerDied","Data":"71de947a8dec65c3ee92f5feadf901bb7d74f27f725fdfb0d808ce3a07822f50"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.276536 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" event={"ID":"e05e03dd-8c4f-423c-bfba-db6d10fef841","Type":"ContainerStarted","Data":"d8b89711705db7b5532bbd0e0103a21a9551fe9b9f807bee73c1c554dc81383a"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.294497 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p4mjk" event={"ID":"5ec32c20-9117-4eca-9b73-a7f8b84677da","Type":"ContainerStarted","Data":"b96444269274e802b8f5aba353bfcf7156bc2e86cdd2fcbd3130f4f86f71ac13"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.319773 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-d8zcz" event={"ID":"d3e2fa00-220e-4811-8b41-644d96b70a46","Type":"ContainerStarted","Data":"1e1b216fe46d6b2ff507fda8255f6644a7f51d023f7bfd4307c0feb95ba11f59"} Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.730079 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.870557 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6d97bc8c5c-g2ds9"] Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.880195 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-config\") pod \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.880288 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-sb\") pod \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.880310 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-swift-storage-0\") pod \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.880360 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-svc\") pod \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.880468 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ltjt\" (UniqueName: \"kubernetes.io/projected/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-kube-api-access-8ltjt\") pod \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.880503 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-nb\") pod \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\" (UID: \"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869\") " Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.915812 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-kube-api-access-8ltjt" (OuterVolumeSpecName: "kube-api-access-8ltjt") pod "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" (UID: "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869"). InnerVolumeSpecName "kube-api-access-8ltjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:11 crc kubenswrapper[4715]: W1203 22:16:11.928742 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda4847d8_02fb_41f4_8a71_0cf1eee4f55b.slice/crio-244765ba472ecb2f2121fb1c87235725e5e924c36cac1540ee33cfc462b432e2 WatchSource:0}: Error finding container 244765ba472ecb2f2121fb1c87235725e5e924c36cac1540ee33cfc462b432e2: Status 404 returned error can't find the container with id 244765ba472ecb2f2121fb1c87235725e5e924c36cac1540ee33cfc462b432e2 Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.940480 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" (UID: "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.993384 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ltjt\" (UniqueName: \"kubernetes.io/projected/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-kube-api-access-8ltjt\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.993426 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:11 crc kubenswrapper[4715]: I1203 22:16:11.996352 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" (UID: "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.016300 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" (UID: "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.019972 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-config" (OuterVolumeSpecName: "config") pod "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" (UID: "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.046089 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" (UID: "b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.095468 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.095497 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.095520 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.095529 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.363874 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"391549ec-5cda-4f77-917c-0b485e1d57bf","Type":"ContainerStarted","Data":"2695a877655a55f58f6a5816c8328729e67bcb6a85d6a89dd42890021cb8a84f"} Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.371488 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" event={"ID":"b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869","Type":"ContainerDied","Data":"4425df45bc1a3c941958509623fa74a8437396092cb60fc60a6494280bcf99e4"} Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.371567 4715 scope.go:117] "RemoveContainer" containerID="24720ddd54bbf1945f48c8116231a134e5ca1e8776a928c63eb38ceecac14c8c" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.373007 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-x92wb" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.384416 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" event={"ID":"e05e03dd-8c4f-423c-bfba-db6d10fef841","Type":"ContainerStarted","Data":"aed158e03dfae84d343df09491228128b814863fd73cf6b514deb3ccef7a9767"} Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.386872 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.389896 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d97bc8c5c-g2ds9" event={"ID":"da4847d8-02fb-41f4-8a71-0cf1eee4f55b","Type":"ContainerStarted","Data":"244765ba472ecb2f2121fb1c87235725e5e924c36cac1540ee33cfc462b432e2"} Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.393591 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec","Type":"ContainerStarted","Data":"8e41c46c571f1f01261fae5162ab6fb89504b85aec2745ca9652f0384f30c9d7"} Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.413624 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" podStartSLOduration=4.41360275 podStartE2EDuration="4.41360275s" podCreationTimestamp="2025-12-03 22:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:16:12.402661187 +0000 UTC m=+2069.145371782" watchObservedRunningTime="2025-12-03 22:16:12.41360275 +0000 UTC m=+2069.156313345" Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.491265 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-x92wb"] Dec 03 22:16:12 crc kubenswrapper[4715]: I1203 22:16:12.537111 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-x92wb"] Dec 03 22:16:13 crc kubenswrapper[4715]: I1203 22:16:13.424114 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec","Type":"ContainerStarted","Data":"3d044d9e8c02229042373a13e1c5938393a12914a01024502c8ccaf261caca35"} Dec 03 22:16:13 crc kubenswrapper[4715]: I1203 22:16:13.424435 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerName="glance-log" containerID="cri-o://8e41c46c571f1f01261fae5162ab6fb89504b85aec2745ca9652f0384f30c9d7" gracePeriod=30 Dec 03 22:16:13 crc kubenswrapper[4715]: I1203 22:16:13.424930 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerName="glance-httpd" containerID="cri-o://3d044d9e8c02229042373a13e1c5938393a12914a01024502c8ccaf261caca35" gracePeriod=30 Dec 03 22:16:13 crc kubenswrapper[4715]: I1203 22:16:13.452170 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.452148402 podStartE2EDuration="5.452148402s" podCreationTimestamp="2025-12-03 22:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:16:13.440872111 +0000 UTC m=+2070.183582726" watchObservedRunningTime="2025-12-03 22:16:13.452148402 +0000 UTC m=+2070.194858997" Dec 03 22:16:13 crc kubenswrapper[4715]: I1203 22:16:13.656177 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" path="/var/lib/kubelet/pods/b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869/volumes" Dec 03 22:16:14 crc kubenswrapper[4715]: I1203 22:16:14.436542 4715 generic.go:334] "Generic (PLEG): container finished" podID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerID="8e41c46c571f1f01261fae5162ab6fb89504b85aec2745ca9652f0384f30c9d7" exitCode=143 Dec 03 22:16:14 crc kubenswrapper[4715]: I1203 22:16:14.436641 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec","Type":"ContainerDied","Data":"8e41c46c571f1f01261fae5162ab6fb89504b85aec2745ca9652f0384f30c9d7"} Dec 03 22:16:15 crc kubenswrapper[4715]: I1203 22:16:15.459640 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"391549ec-5cda-4f77-917c-0b485e1d57bf","Type":"ContainerStarted","Data":"c8f680048e38eb0e1f98013c689bb324de46fe5e1d8acd5f7b64c3526203a274"} Dec 03 22:16:15 crc kubenswrapper[4715]: I1203 22:16:15.459760 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerName="glance-log" containerID="cri-o://2695a877655a55f58f6a5816c8328729e67bcb6a85d6a89dd42890021cb8a84f" gracePeriod=30 Dec 03 22:16:15 crc kubenswrapper[4715]: I1203 22:16:15.459838 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerName="glance-httpd" containerID="cri-o://c8f680048e38eb0e1f98013c689bb324de46fe5e1d8acd5f7b64c3526203a274" gracePeriod=30 Dec 03 22:16:15 crc kubenswrapper[4715]: I1203 22:16:15.466398 4715 generic.go:334] "Generic (PLEG): container finished" podID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerID="3d044d9e8c02229042373a13e1c5938393a12914a01024502c8ccaf261caca35" exitCode=0 Dec 03 22:16:15 crc kubenswrapper[4715]: I1203 22:16:15.466444 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec","Type":"ContainerDied","Data":"3d044d9e8c02229042373a13e1c5938393a12914a01024502c8ccaf261caca35"} Dec 03 22:16:15 crc kubenswrapper[4715]: I1203 22:16:15.488289 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=7.48826881 podStartE2EDuration="7.48826881s" podCreationTimestamp="2025-12-03 22:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:16:15.478459358 +0000 UTC m=+2072.221169953" watchObservedRunningTime="2025-12-03 22:16:15.48826881 +0000 UTC m=+2072.230979405" Dec 03 22:16:16 crc kubenswrapper[4715]: I1203 22:16:16.490920 4715 generic.go:334] "Generic (PLEG): container finished" podID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerID="c8f680048e38eb0e1f98013c689bb324de46fe5e1d8acd5f7b64c3526203a274" exitCode=0 Dec 03 22:16:16 crc kubenswrapper[4715]: I1203 22:16:16.491295 4715 generic.go:334] "Generic (PLEG): container finished" podID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerID="2695a877655a55f58f6a5816c8328729e67bcb6a85d6a89dd42890021cb8a84f" exitCode=143 Dec 03 22:16:16 crc kubenswrapper[4715]: I1203 22:16:16.491009 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"391549ec-5cda-4f77-917c-0b485e1d57bf","Type":"ContainerDied","Data":"c8f680048e38eb0e1f98013c689bb324de46fe5e1d8acd5f7b64c3526203a274"} Dec 03 22:16:16 crc kubenswrapper[4715]: I1203 22:16:16.491379 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"391549ec-5cda-4f77-917c-0b485e1d57bf","Type":"ContainerDied","Data":"2695a877655a55f58f6a5816c8328729e67bcb6a85d6a89dd42890021cb8a84f"} Dec 03 22:16:16 crc kubenswrapper[4715]: I1203 22:16:16.495424 4715 generic.go:334] "Generic (PLEG): container finished" podID="5a7bfab0-bbd0-4773-a92a-54449b1f486b" containerID="a5b0d903b1f711b99655d67b5066c14cddb95948e7c959d7124c9576077583ba" exitCode=0 Dec 03 22:16:16 crc kubenswrapper[4715]: I1203 22:16:16.495541 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pgkfz" event={"ID":"5a7bfab0-bbd0-4773-a92a-54449b1f486b","Type":"ContainerDied","Data":"a5b0d903b1f711b99655d67b5066c14cddb95948e7c959d7124c9576077583ba"} Dec 03 22:16:17 crc kubenswrapper[4715]: I1203 22:16:17.909064 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-868658b5-6jkj6"] Dec 03 22:16:17 crc kubenswrapper[4715]: I1203 22:16:17.982515 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74cd84ccf4-k9bcn"] Dec 03 22:16:17 crc kubenswrapper[4715]: E1203 22:16:17.982968 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" containerName="init" Dec 03 22:16:17 crc kubenswrapper[4715]: I1203 22:16:17.982982 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" containerName="init" Dec 03 22:16:17 crc kubenswrapper[4715]: I1203 22:16:17.983171 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2cab2e3-a83d-4069-a5ba-5c5e5c7c0869" containerName="init" Dec 03 22:16:17 crc kubenswrapper[4715]: I1203 22:16:17.984040 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:17 crc kubenswrapper[4715]: I1203 22:16:17.989965 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.008370 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74cd84ccf4-k9bcn"] Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.093232 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-combined-ca-bundle\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.093313 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-config-data\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.093389 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-tls-certs\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.093407 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sp89\" (UniqueName: \"kubernetes.io/projected/ae09bc5f-d3e1-48c3-8145-611fd9827870-kube-api-access-5sp89\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.093470 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-secret-key\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.093538 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-scripts\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.093569 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae09bc5f-d3e1-48c3-8145-611fd9827870-logs\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.173854 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d97bc8c5c-g2ds9"] Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.181784 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-656478594d-5tbrm"] Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.183255 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.195410 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-combined-ca-bundle\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.195493 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-config-data\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.195601 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-tls-certs\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.195627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sp89\" (UniqueName: \"kubernetes.io/projected/ae09bc5f-d3e1-48c3-8145-611fd9827870-kube-api-access-5sp89\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.195680 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-secret-key\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.195722 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-scripts\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.195755 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae09bc5f-d3e1-48c3-8145-611fd9827870-logs\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.196431 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae09bc5f-d3e1-48c3-8145-611fd9827870-logs\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.198572 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-config-data\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.200617 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-scripts\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.209114 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-656478594d-5tbrm"] Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.210933 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-tls-certs\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.216237 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-secret-key\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.219438 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-combined-ca-bundle\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.238762 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sp89\" (UniqueName: \"kubernetes.io/projected/ae09bc5f-d3e1-48c3-8145-611fd9827870-kube-api-access-5sp89\") pod \"horizon-74cd84ccf4-k9bcn\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.298342 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-horizon-tls-certs\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.298845 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c7cfacf-a283-45d3-982d-b28e62a9146b-config-data\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.299147 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gw7d\" (UniqueName: \"kubernetes.io/projected/6c7cfacf-a283-45d3-982d-b28e62a9146b-kube-api-access-2gw7d\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.299184 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c7cfacf-a283-45d3-982d-b28e62a9146b-logs\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.299236 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-combined-ca-bundle\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.299319 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c7cfacf-a283-45d3-982d-b28e62a9146b-scripts\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.299409 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-horizon-secret-key\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.313660 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.401557 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-horizon-tls-certs\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.401645 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c7cfacf-a283-45d3-982d-b28e62a9146b-config-data\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.401695 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gw7d\" (UniqueName: \"kubernetes.io/projected/6c7cfacf-a283-45d3-982d-b28e62a9146b-kube-api-access-2gw7d\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.401716 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c7cfacf-a283-45d3-982d-b28e62a9146b-logs\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.401732 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-combined-ca-bundle\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.401748 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c7cfacf-a283-45d3-982d-b28e62a9146b-scripts\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.401772 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-horizon-secret-key\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.403477 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6c7cfacf-a283-45d3-982d-b28e62a9146b-logs\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.403829 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6c7cfacf-a283-45d3-982d-b28e62a9146b-scripts\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.404589 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6c7cfacf-a283-45d3-982d-b28e62a9146b-config-data\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.405841 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-horizon-tls-certs\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.407422 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-horizon-secret-key\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.409837 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c7cfacf-a283-45d3-982d-b28e62a9146b-combined-ca-bundle\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.420685 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gw7d\" (UniqueName: \"kubernetes.io/projected/6c7cfacf-a283-45d3-982d-b28e62a9146b-kube-api-access-2gw7d\") pod \"horizon-656478594d-5tbrm\" (UID: \"6c7cfacf-a283-45d3-982d-b28e62a9146b\") " pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:18 crc kubenswrapper[4715]: I1203 22:16:18.511362 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:19 crc kubenswrapper[4715]: I1203 22:16:19.439708 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:16:19 crc kubenswrapper[4715]: I1203 22:16:19.503094 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-f5kgj"] Dec 03 22:16:19 crc kubenswrapper[4715]: I1203 22:16:19.505618 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="dnsmasq-dns" containerID="cri-o://c7755f1a98a28567ccae06ca7899178a25b184da6e2f67de8f9a4ce23cd2b684" gracePeriod=10 Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.601927 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.602979 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec","Type":"ContainerDied","Data":"6e90802a0e9412beed9c3f0b46f3b07af8b09aa86b0f3961e061070013f4b613"} Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.603028 4715 scope.go:117] "RemoveContainer" containerID="3d044d9e8c02229042373a13e1c5938393a12914a01024502c8ccaf261caca35" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.610399 4715 generic.go:334] "Generic (PLEG): container finished" podID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerID="c7755f1a98a28567ccae06ca7899178a25b184da6e2f67de8f9a4ce23cd2b684" exitCode=0 Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.610433 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" event={"ID":"74cdaa65-5773-4ff8-b84d-31100df6f576","Type":"ContainerDied","Data":"c7755f1a98a28567ccae06ca7899178a25b184da6e2f67de8f9a4ce23cd2b684"} Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.755186 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-public-tls-certs\") pod \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.755244 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vklm5\" (UniqueName: \"kubernetes.io/projected/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-kube-api-access-vklm5\") pod \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.755306 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.755339 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-logs\") pod \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.755426 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-config-data\") pod \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.755481 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-combined-ca-bundle\") pod \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.755592 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-httpd-run\") pod \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.755657 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-scripts\") pod \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\" (UID: \"0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec\") " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.757056 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" (UID: "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.757164 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-logs" (OuterVolumeSpecName: "logs") pod "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" (UID: "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.761178 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-scripts" (OuterVolumeSpecName: "scripts") pod "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" (UID: "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.762074 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-kube-api-access-vklm5" (OuterVolumeSpecName: "kube-api-access-vklm5") pod "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" (UID: "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec"). InnerVolumeSpecName "kube-api-access-vklm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.786575 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" (UID: "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.787115 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" (UID: "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.803707 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" (UID: "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.809073 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-config-data" (OuterVolumeSpecName: "config-data") pod "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" (UID: "0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.857447 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.857479 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.857491 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.857516 4715 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.857527 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vklm5\" (UniqueName: \"kubernetes.io/projected/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-kube-api-access-vklm5\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.857562 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.857572 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.857583 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.874275 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 22:16:20 crc kubenswrapper[4715]: I1203 22:16:20.959623 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.621219 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.667936 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.701604 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.718263 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.725874 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:21 crc kubenswrapper[4715]: E1203 22:16:21.727140 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerName="glance-log" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.727163 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerName="glance-log" Dec 03 22:16:21 crc kubenswrapper[4715]: E1203 22:16:21.727194 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerName="glance-httpd" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.727200 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerName="glance-httpd" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.727376 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerName="glance-log" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.727405 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" containerName="glance-httpd" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.728495 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.731466 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.732336 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.738146 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.883483 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.883554 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-logs\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.883572 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc8qp\" (UniqueName: \"kubernetes.io/projected/6143cec8-8a39-443c-b850-109e24ac1200-kube-api-access-pc8qp\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.883601 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-scripts\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.883628 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.883661 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.883701 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.883732 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-config-data\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985465 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-logs\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985649 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc8qp\" (UniqueName: \"kubernetes.io/projected/6143cec8-8a39-443c-b850-109e24ac1200-kube-api-access-pc8qp\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985708 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-scripts\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985765 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985834 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985922 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985962 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.985989 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-config-data\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.986186 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.986990 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-logs\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.991710 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.992202 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-config-data\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:21 crc kubenswrapper[4715]: I1203 22:16:21.993008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-scripts\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:22 crc kubenswrapper[4715]: I1203 22:16:22.000930 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:22 crc kubenswrapper[4715]: I1203 22:16:22.005751 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc8qp\" (UniqueName: \"kubernetes.io/projected/6143cec8-8a39-443c-b850-109e24ac1200-kube-api-access-pc8qp\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:22 crc kubenswrapper[4715]: I1203 22:16:22.012944 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " pod="openstack/glance-default-external-api-0" Dec 03 22:16:22 crc kubenswrapper[4715]: I1203 22:16:22.064255 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:16:24 crc kubenswrapper[4715]: I1203 22:16:23.649237 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec" path="/var/lib/kubelet/pods/0d62a7f8-9c72-4eed-9d5c-c61dfc8cf8ec/volumes" Dec 03 22:16:25 crc kubenswrapper[4715]: E1203 22:16:25.633757 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 22:16:25 crc kubenswrapper[4715]: E1203 22:16:25.634417 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d8h5d8h5b8h599h9dhc6h579h5f8h5b5hf7h558h8dh4h68ch5d9h65hf4h596h698h76h58fh584h64ch76h5ddh5f4h65fh55ch65fh646h8bh645q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-57rms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-55967db74f-69gh7_openstack(98853dba-6c7a-4ed5-afa4-3696e1e1eee6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:16:25 crc kubenswrapper[4715]: E1203 22:16:25.637374 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-55967db74f-69gh7" podUID="98853dba-6c7a-4ed5-afa4-3696e1e1eee6" Dec 03 22:16:26 crc kubenswrapper[4715]: I1203 22:16:26.666697 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Dec 03 22:16:31 crc kubenswrapper[4715]: E1203 22:16:31.281218 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 22:16:31 crc kubenswrapper[4715]: E1203 22:16:31.282523 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n575h4h574h67bh654h689h95h5c5h646h65dh599h559hbh96h66fh65dhb8h65h6ch66bhc9hb8h678hddhdch67fh594hd7h78h5d8h5f9h77q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n4ch7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-868658b5-6jkj6_openstack(dfb98490-b33a-46a6-9f7d-668862bd0430): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:16:31 crc kubenswrapper[4715]: E1203 22:16:31.287463 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-868658b5-6jkj6" podUID="dfb98490-b33a-46a6-9f7d-668862bd0430" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.389292 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:31 crc kubenswrapper[4715]: E1203 22:16:31.458185 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Dec 03 22:16:31 crc kubenswrapper[4715]: E1203 22:16:31.458427 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n67h7hbch7bhc4h57dh545h556h699h5f8h9hfbh664h5ddh5cdh554hb7h549h9dh6bh577h58h575h548hfbh5ch654h58dh66fh55ch598hbbq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9dxjs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6d97bc8c5c-g2ds9_openstack(da4847d8-02fb-41f4-8a71-0cf1eee4f55b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:16:31 crc kubenswrapper[4715]: E1203 22:16:31.461166 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-6d97bc8c5c-g2ds9" podUID="da4847d8-02fb-41f4-8a71-0cf1eee4f55b" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.504180 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-credential-keys\") pod \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.504256 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-combined-ca-bundle\") pod \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.504773 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks8gn\" (UniqueName: \"kubernetes.io/projected/5a7bfab0-bbd0-4773-a92a-54449b1f486b-kube-api-access-ks8gn\") pod \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.504828 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-scripts\") pod \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.505150 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-config-data\") pod \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.505352 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-fernet-keys\") pod \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\" (UID: \"5a7bfab0-bbd0-4773-a92a-54449b1f486b\") " Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.512881 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5a7bfab0-bbd0-4773-a92a-54449b1f486b" (UID: "5a7bfab0-bbd0-4773-a92a-54449b1f486b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.512991 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a7bfab0-bbd0-4773-a92a-54449b1f486b-kube-api-access-ks8gn" (OuterVolumeSpecName: "kube-api-access-ks8gn") pod "5a7bfab0-bbd0-4773-a92a-54449b1f486b" (UID: "5a7bfab0-bbd0-4773-a92a-54449b1f486b"). InnerVolumeSpecName "kube-api-access-ks8gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.519619 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-scripts" (OuterVolumeSpecName: "scripts") pod "5a7bfab0-bbd0-4773-a92a-54449b1f486b" (UID: "5a7bfab0-bbd0-4773-a92a-54449b1f486b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.527209 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5a7bfab0-bbd0-4773-a92a-54449b1f486b" (UID: "5a7bfab0-bbd0-4773-a92a-54449b1f486b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.539612 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a7bfab0-bbd0-4773-a92a-54449b1f486b" (UID: "5a7bfab0-bbd0-4773-a92a-54449b1f486b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.552164 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-config-data" (OuterVolumeSpecName: "config-data") pod "5a7bfab0-bbd0-4773-a92a-54449b1f486b" (UID: "5a7bfab0-bbd0-4773-a92a-54449b1f486b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.611844 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks8gn\" (UniqueName: \"kubernetes.io/projected/5a7bfab0-bbd0-4773-a92a-54449b1f486b-kube-api-access-ks8gn\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.611891 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.611901 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.611935 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.611944 4715 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.611953 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7bfab0-bbd0-4773-a92a-54449b1f486b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.747890 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-pgkfz" Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.748596 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-pgkfz" event={"ID":"5a7bfab0-bbd0-4773-a92a-54449b1f486b","Type":"ContainerDied","Data":"c4a39a7dad139e08366ecbf387d73e370b5ef182fc0505f732aeb4cdce19a27a"} Dec 03 22:16:31 crc kubenswrapper[4715]: I1203 22:16:31.748622 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4a39a7dad139e08366ecbf387d73e370b5ef182fc0505f732aeb4cdce19a27a" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.501470 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-pgkfz"] Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.517422 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-pgkfz"] Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.607455 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-kp478"] Dec 03 22:16:32 crc kubenswrapper[4715]: E1203 22:16:32.608380 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a7bfab0-bbd0-4773-a92a-54449b1f486b" containerName="keystone-bootstrap" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.608427 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a7bfab0-bbd0-4773-a92a-54449b1f486b" containerName="keystone-bootstrap" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.609287 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a7bfab0-bbd0-4773-a92a-54449b1f486b" containerName="keystone-bootstrap" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.610743 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.613907 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.614709 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9bcdn" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.620028 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.622822 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.623127 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.629132 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kp478"] Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.701381 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-config-data\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.701698 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-fernet-keys\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.701850 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-combined-ca-bundle\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.701980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-scripts\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.702190 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gjjcm\" (UniqueName: \"kubernetes.io/projected/60edd1a1-9b8a-4d7d-977b-11c834922ca1-kube-api-access-gjjcm\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.702278 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-credential-keys\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.804493 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-config-data\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.805586 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-fernet-keys\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.805632 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-combined-ca-bundle\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.805865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-scripts\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.805915 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gjjcm\" (UniqueName: \"kubernetes.io/projected/60edd1a1-9b8a-4d7d-977b-11c834922ca1-kube-api-access-gjjcm\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.805949 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-credential-keys\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.814186 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-credential-keys\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.816579 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-combined-ca-bundle\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.816722 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-scripts\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.816880 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-fernet-keys\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.817567 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-config-data\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.838172 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gjjcm\" (UniqueName: \"kubernetes.io/projected/60edd1a1-9b8a-4d7d-977b-11c834922ca1-kube-api-access-gjjcm\") pod \"keystone-bootstrap-kp478\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:32 crc kubenswrapper[4715]: I1203 22:16:32.946042 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kp478" Dec 03 22:16:33 crc kubenswrapper[4715]: E1203 22:16:33.050048 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Dec 03 22:16:33 crc kubenswrapper[4715]: E1203 22:16:33.050235 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dsskp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-6ngmv_openstack(2eba90e4-4991-4260-9e8e-2a3a5412b264): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:16:33 crc kubenswrapper[4715]: E1203 22:16:33.051888 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-6ngmv" podUID="2eba90e4-4991-4260-9e8e-2a3a5412b264" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.118379 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.131857 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.212923 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-svc\") pod \"74cdaa65-5773-4ff8-b84d-31100df6f576\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.213047 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dfb98490-b33a-46a6-9f7d-668862bd0430-horizon-secret-key\") pod \"dfb98490-b33a-46a6-9f7d-668862bd0430\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.213165 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-nb\") pod \"74cdaa65-5773-4ff8-b84d-31100df6f576\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.213695 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb98490-b33a-46a6-9f7d-668862bd0430-logs\") pod \"dfb98490-b33a-46a6-9f7d-668862bd0430\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.213748 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-swift-storage-0\") pod \"74cdaa65-5773-4ff8-b84d-31100df6f576\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.213790 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-sb\") pod \"74cdaa65-5773-4ff8-b84d-31100df6f576\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.213848 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4ch7\" (UniqueName: \"kubernetes.io/projected/dfb98490-b33a-46a6-9f7d-668862bd0430-kube-api-access-n4ch7\") pod \"dfb98490-b33a-46a6-9f7d-668862bd0430\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.213931 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-config-data\") pod \"dfb98490-b33a-46a6-9f7d-668862bd0430\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.214021 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcxpp\" (UniqueName: \"kubernetes.io/projected/74cdaa65-5773-4ff8-b84d-31100df6f576-kube-api-access-dcxpp\") pod \"74cdaa65-5773-4ff8-b84d-31100df6f576\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.214139 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-scripts\") pod \"dfb98490-b33a-46a6-9f7d-668862bd0430\" (UID: \"dfb98490-b33a-46a6-9f7d-668862bd0430\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.214187 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-config\") pod \"74cdaa65-5773-4ff8-b84d-31100df6f576\" (UID: \"74cdaa65-5773-4ff8-b84d-31100df6f576\") " Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.218934 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfb98490-b33a-46a6-9f7d-668862bd0430-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "dfb98490-b33a-46a6-9f7d-668862bd0430" (UID: "dfb98490-b33a-46a6-9f7d-668862bd0430"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.218994 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-scripts" (OuterVolumeSpecName: "scripts") pod "dfb98490-b33a-46a6-9f7d-668862bd0430" (UID: "dfb98490-b33a-46a6-9f7d-668862bd0430"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.219249 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfb98490-b33a-46a6-9f7d-668862bd0430-logs" (OuterVolumeSpecName: "logs") pod "dfb98490-b33a-46a6-9f7d-668862bd0430" (UID: "dfb98490-b33a-46a6-9f7d-668862bd0430"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.219591 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-config-data" (OuterVolumeSpecName: "config-data") pod "dfb98490-b33a-46a6-9f7d-668862bd0430" (UID: "dfb98490-b33a-46a6-9f7d-668862bd0430"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.219904 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfb98490-b33a-46a6-9f7d-668862bd0430-kube-api-access-n4ch7" (OuterVolumeSpecName: "kube-api-access-n4ch7") pod "dfb98490-b33a-46a6-9f7d-668862bd0430" (UID: "dfb98490-b33a-46a6-9f7d-668862bd0430"). InnerVolumeSpecName "kube-api-access-n4ch7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.226821 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74cdaa65-5773-4ff8-b84d-31100df6f576-kube-api-access-dcxpp" (OuterVolumeSpecName: "kube-api-access-dcxpp") pod "74cdaa65-5773-4ff8-b84d-31100df6f576" (UID: "74cdaa65-5773-4ff8-b84d-31100df6f576"). InnerVolumeSpecName "kube-api-access-dcxpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.262793 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "74cdaa65-5773-4ff8-b84d-31100df6f576" (UID: "74cdaa65-5773-4ff8-b84d-31100df6f576"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.265834 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-config" (OuterVolumeSpecName: "config") pod "74cdaa65-5773-4ff8-b84d-31100df6f576" (UID: "74cdaa65-5773-4ff8-b84d-31100df6f576"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.282495 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "74cdaa65-5773-4ff8-b84d-31100df6f576" (UID: "74cdaa65-5773-4ff8-b84d-31100df6f576"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.283720 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "74cdaa65-5773-4ff8-b84d-31100df6f576" (UID: "74cdaa65-5773-4ff8-b84d-31100df6f576"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.288164 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "74cdaa65-5773-4ff8-b84d-31100df6f576" (UID: "74cdaa65-5773-4ff8-b84d-31100df6f576"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.315932 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.315964 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/dfb98490-b33a-46a6-9f7d-668862bd0430-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.315979 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.315988 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dfb98490-b33a-46a6-9f7d-668862bd0430-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.315996 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.316005 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.316013 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4ch7\" (UniqueName: \"kubernetes.io/projected/dfb98490-b33a-46a6-9f7d-668862bd0430-kube-api-access-n4ch7\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.316022 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.316031 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcxpp\" (UniqueName: \"kubernetes.io/projected/74cdaa65-5773-4ff8-b84d-31100df6f576-kube-api-access-dcxpp\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.316040 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dfb98490-b33a-46a6-9f7d-668862bd0430-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.316049 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74cdaa65-5773-4ff8-b84d-31100df6f576-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.651315 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a7bfab0-bbd0-4773-a92a-54449b1f486b" path="/var/lib/kubelet/pods/5a7bfab0-bbd0-4773-a92a-54449b1f486b/volumes" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.769715 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-868658b5-6jkj6" event={"ID":"dfb98490-b33a-46a6-9f7d-668862bd0430","Type":"ContainerDied","Data":"c9427c5d66d898d1c1c37665df0f0a69ba873b7702ae617452dc51c42ed83ebe"} Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.769734 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-868658b5-6jkj6" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.773348 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.773371 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" event={"ID":"74cdaa65-5773-4ff8-b84d-31100df6f576","Type":"ContainerDied","Data":"3b4e150e07844647a418fca22372f5889863a016fbb1141e1d0f910e45d9b46a"} Dec 03 22:16:33 crc kubenswrapper[4715]: E1203 22:16:33.775551 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-6ngmv" podUID="2eba90e4-4991-4260-9e8e-2a3a5412b264" Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.815434 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-f5kgj"] Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.824574 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-f5kgj"] Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.872285 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-868658b5-6jkj6"] Dec 03 22:16:33 crc kubenswrapper[4715]: I1203 22:16:33.880461 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-868658b5-6jkj6"] Dec 03 22:16:35 crc kubenswrapper[4715]: I1203 22:16:35.646381 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" path="/var/lib/kubelet/pods/74cdaa65-5773-4ff8-b84d-31100df6f576/volumes" Dec 03 22:16:35 crc kubenswrapper[4715]: I1203 22:16:35.648052 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfb98490-b33a-46a6-9f7d-668862bd0430" path="/var/lib/kubelet/pods/dfb98490-b33a-46a6-9f7d-668862bd0430/volumes" Dec 03 22:16:36 crc kubenswrapper[4715]: I1203 22:16:36.667603 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-f5kgj" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Dec 03 22:16:39 crc kubenswrapper[4715]: I1203 22:16:39.515700 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:39 crc kubenswrapper[4715]: I1203 22:16:39.516315 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.474592 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.608852 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-config-data\") pod \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.609000 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-scripts\") pod \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.609171 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57rms\" (UniqueName: \"kubernetes.io/projected/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-kube-api-access-57rms\") pod \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.609323 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-logs\") pod \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.609540 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-horizon-secret-key\") pod \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\" (UID: \"98853dba-6c7a-4ed5-afa4-3696e1e1eee6\") " Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.609649 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-scripts" (OuterVolumeSpecName: "scripts") pod "98853dba-6c7a-4ed5-afa4-3696e1e1eee6" (UID: "98853dba-6c7a-4ed5-afa4-3696e1e1eee6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.609783 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-logs" (OuterVolumeSpecName: "logs") pod "98853dba-6c7a-4ed5-afa4-3696e1e1eee6" (UID: "98853dba-6c7a-4ed5-afa4-3696e1e1eee6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.610321 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-config-data" (OuterVolumeSpecName: "config-data") pod "98853dba-6c7a-4ed5-afa4-3696e1e1eee6" (UID: "98853dba-6c7a-4ed5-afa4-3696e1e1eee6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.610415 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.610455 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.616916 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "98853dba-6c7a-4ed5-afa4-3696e1e1eee6" (UID: "98853dba-6c7a-4ed5-afa4-3696e1e1eee6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.617337 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-kube-api-access-57rms" (OuterVolumeSpecName: "kube-api-access-57rms") pod "98853dba-6c7a-4ed5-afa4-3696e1e1eee6" (UID: "98853dba-6c7a-4ed5-afa4-3696e1e1eee6"). InnerVolumeSpecName "kube-api-access-57rms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.714923 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.714969 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.714992 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57rms\" (UniqueName: \"kubernetes.io/projected/98853dba-6c7a-4ed5-afa4-3696e1e1eee6-kube-api-access-57rms\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.911427 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55967db74f-69gh7" event={"ID":"98853dba-6c7a-4ed5-afa4-3696e1e1eee6","Type":"ContainerDied","Data":"84225bacf08c0f9534733f5ea0c81fb190f90ee005f8b6c9ccd62ff18b9a831c"} Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.911540 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55967db74f-69gh7" Dec 03 22:16:46 crc kubenswrapper[4715]: I1203 22:16:46.991178 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55967db74f-69gh7"] Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.000117 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-55967db74f-69gh7"] Dec 03 22:16:47 crc kubenswrapper[4715]: E1203 22:16:47.165591 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 22:16:47 crc kubenswrapper[4715]: E1203 22:16:47.165845 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n56ch5d5h689hffh645h54dh55dh54h5bch557h59ch574h5bdh68dh677h699h67bh9h7ch67ch548h646h5hb8h56dh5b6h5bh64dh5fch576h6bhf9q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2lp2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(72b03167-1df1-47c3-94c7-e0f13b65f972): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.193957 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.290825 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.301141 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426253 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-horizon-secret-key\") pod \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426345 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-combined-ca-bundle\") pod \"391549ec-5cda-4f77-917c-0b485e1d57bf\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426386 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dxjs\" (UniqueName: \"kubernetes.io/projected/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-kube-api-access-9dxjs\") pod \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426405 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxfvz\" (UniqueName: \"kubernetes.io/projected/391549ec-5cda-4f77-917c-0b485e1d57bf-kube-api-access-gxfvz\") pod \"391549ec-5cda-4f77-917c-0b485e1d57bf\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426444 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-config-data\") pod \"391549ec-5cda-4f77-917c-0b485e1d57bf\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426473 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-internal-tls-certs\") pod \"391549ec-5cda-4f77-917c-0b485e1d57bf\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426526 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-config-data\") pod \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426570 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"391549ec-5cda-4f77-917c-0b485e1d57bf\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426615 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-scripts\") pod \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426648 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-logs\") pod \"391549ec-5cda-4f77-917c-0b485e1d57bf\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-httpd-run\") pod \"391549ec-5cda-4f77-917c-0b485e1d57bf\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426688 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-scripts\") pod \"391549ec-5cda-4f77-917c-0b485e1d57bf\" (UID: \"391549ec-5cda-4f77-917c-0b485e1d57bf\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.426774 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-logs\") pod \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\" (UID: \"da4847d8-02fb-41f4-8a71-0cf1eee4f55b\") " Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.427408 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-logs" (OuterVolumeSpecName: "logs") pod "da4847d8-02fb-41f4-8a71-0cf1eee4f55b" (UID: "da4847d8-02fb-41f4-8a71-0cf1eee4f55b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.427463 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-logs" (OuterVolumeSpecName: "logs") pod "391549ec-5cda-4f77-917c-0b485e1d57bf" (UID: "391549ec-5cda-4f77-917c-0b485e1d57bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.427746 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "391549ec-5cda-4f77-917c-0b485e1d57bf" (UID: "391549ec-5cda-4f77-917c-0b485e1d57bf"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.428223 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-scripts" (OuterVolumeSpecName: "scripts") pod "da4847d8-02fb-41f4-8a71-0cf1eee4f55b" (UID: "da4847d8-02fb-41f4-8a71-0cf1eee4f55b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.428415 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-config-data" (OuterVolumeSpecName: "config-data") pod "da4847d8-02fb-41f4-8a71-0cf1eee4f55b" (UID: "da4847d8-02fb-41f4-8a71-0cf1eee4f55b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:16:47 crc kubenswrapper[4715]: I1203 22:16:47.433822 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "391549ec-5cda-4f77-917c-0b485e1d57bf" (UID: "391549ec-5cda-4f77-917c-0b485e1d57bf"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.433890 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-kube-api-access-9dxjs" (OuterVolumeSpecName: "kube-api-access-9dxjs") pod "da4847d8-02fb-41f4-8a71-0cf1eee4f55b" (UID: "da4847d8-02fb-41f4-8a71-0cf1eee4f55b"). InnerVolumeSpecName "kube-api-access-9dxjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.435096 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/391549ec-5cda-4f77-917c-0b485e1d57bf-kube-api-access-gxfvz" (OuterVolumeSpecName: "kube-api-access-gxfvz") pod "391549ec-5cda-4f77-917c-0b485e1d57bf" (UID: "391549ec-5cda-4f77-917c-0b485e1d57bf"). InnerVolumeSpecName "kube-api-access-gxfvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.435301 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "da4847d8-02fb-41f4-8a71-0cf1eee4f55b" (UID: "da4847d8-02fb-41f4-8a71-0cf1eee4f55b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.439609 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-scripts" (OuterVolumeSpecName: "scripts") pod "391549ec-5cda-4f77-917c-0b485e1d57bf" (UID: "391549ec-5cda-4f77-917c-0b485e1d57bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.451407 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "391549ec-5cda-4f77-917c-0b485e1d57bf" (UID: "391549ec-5cda-4f77-917c-0b485e1d57bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.485681 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "391549ec-5cda-4f77-917c-0b485e1d57bf" (UID: "391549ec-5cda-4f77-917c-0b485e1d57bf"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.504425 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-config-data" (OuterVolumeSpecName: "config-data") pod "391549ec-5cda-4f77-917c-0b485e1d57bf" (UID: "391549ec-5cda-4f77-917c-0b485e1d57bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528796 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528827 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dxjs\" (UniqueName: \"kubernetes.io/projected/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-kube-api-access-9dxjs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528843 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxfvz\" (UniqueName: \"kubernetes.io/projected/391549ec-5cda-4f77-917c-0b485e1d57bf-kube-api-access-gxfvz\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528855 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528866 4715 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528877 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528914 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528927 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528940 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528951 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/391549ec-5cda-4f77-917c-0b485e1d57bf-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528960 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/391549ec-5cda-4f77-917c-0b485e1d57bf-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528971 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.528981 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/da4847d8-02fb-41f4-8a71-0cf1eee4f55b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.551699 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.630492 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.646966 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98853dba-6c7a-4ed5-afa4-3696e1e1eee6" path="/var/lib/kubelet/pods/98853dba-6c7a-4ed5-afa4-3696e1e1eee6/volumes" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.923471 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"391549ec-5cda-4f77-917c-0b485e1d57bf","Type":"ContainerDied","Data":"8858e04e32485c764a9ff25c65cfc910575f2df7dac66cff0e4171abd71c0fe8"} Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.923619 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.925124 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6d97bc8c5c-g2ds9" event={"ID":"da4847d8-02fb-41f4-8a71-0cf1eee4f55b","Type":"ContainerDied","Data":"244765ba472ecb2f2121fb1c87235725e5e924c36cac1540ee33cfc462b432e2"} Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.925180 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6d97bc8c5c-g2ds9" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:47.978265 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.004573 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.023417 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:48 crc kubenswrapper[4715]: E1203 22:16:48.024138 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="init" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.024159 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="init" Dec 03 22:16:48 crc kubenswrapper[4715]: E1203 22:16:48.024185 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerName="glance-httpd" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.024198 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerName="glance-httpd" Dec 03 22:16:48 crc kubenswrapper[4715]: E1203 22:16:48.024233 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerName="glance-log" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.024245 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerName="glance-log" Dec 03 22:16:48 crc kubenswrapper[4715]: E1203 22:16:48.024283 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="dnsmasq-dns" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.024296 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="dnsmasq-dns" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.024645 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="74cdaa65-5773-4ff8-b84d-31100df6f576" containerName="dnsmasq-dns" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.024675 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerName="glance-log" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.024705 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" containerName="glance-httpd" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.026216 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.028977 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.029619 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.044541 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6d97bc8c5c-g2ds9"] Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.056397 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6d97bc8c5c-g2ds9"] Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.066980 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.138852 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.138905 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.138937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-logs\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.138955 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.138978 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.139034 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.139059 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.139098 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8lw5\" (UniqueName: \"kubernetes.io/projected/279fe539-9749-4528-b219-df6323816ea7-kube-api-access-p8lw5\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.240897 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.240998 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.241061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-logs\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.241110 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.241158 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.241272 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.241311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.241398 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8lw5\" (UniqueName: \"kubernetes.io/projected/279fe539-9749-4528-b219-df6323816ea7-kube-api-access-p8lw5\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.242004 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.242309 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.242381 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-logs\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.246305 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.247817 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.254020 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.254835 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.257588 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8lw5\" (UniqueName: \"kubernetes.io/projected/279fe539-9749-4528-b219-df6323816ea7-kube-api-access-p8lw5\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.264862 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.355987 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.514749 4715 scope.go:117] "RemoveContainer" containerID="8e41c46c571f1f01261fae5162ab6fb89504b85aec2745ca9652f0384f30c9d7" Dec 03 22:16:48 crc kubenswrapper[4715]: E1203 22:16:48.580013 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 22:16:48 crc kubenswrapper[4715]: E1203 22:16:48.580260 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9t27x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-p4mjk_openstack(5ec32c20-9117-4eca-9b73-a7f8b84677da): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:16:48 crc kubenswrapper[4715]: E1203 22:16:48.581753 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-p4mjk" podUID="5ec32c20-9117-4eca-9b73-a7f8b84677da" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.818080 4715 scope.go:117] "RemoveContainer" containerID="c7755f1a98a28567ccae06ca7899178a25b184da6e2f67de8f9a4ce23cd2b684" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.844200 4715 scope.go:117] "RemoveContainer" containerID="906b471218ec478389c8bf75c00fd4fe476fbf74018532f45fc55547fc837028" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.863713 4715 scope.go:117] "RemoveContainer" containerID="c8f680048e38eb0e1f98013c689bb324de46fe5e1d8acd5f7b64c3526203a274" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.883578 4715 scope.go:117] "RemoveContainer" containerID="2695a877655a55f58f6a5816c8328729e67bcb6a85d6a89dd42890021cb8a84f" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.953450 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-d8zcz" event={"ID":"d3e2fa00-220e-4811-8b41-644d96b70a46","Type":"ContainerStarted","Data":"3fa789e13ceb6805ffdd0b41900fc86eed3857e22eaf5896774011c906deb354"} Dec 03 22:16:48 crc kubenswrapper[4715]: E1203 22:16:48.958721 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-p4mjk" podUID="5ec32c20-9117-4eca-9b73-a7f8b84677da" Dec 03 22:16:48 crc kubenswrapper[4715]: I1203 22:16:48.982762 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-d8zcz" podStartSLOduration=2.6850986089999997 podStartE2EDuration="40.98273581s" podCreationTimestamp="2025-12-03 22:16:08 +0000 UTC" firstStartedPulling="2025-12-03 22:16:10.241413884 +0000 UTC m=+2066.984124479" lastFinishedPulling="2025-12-03 22:16:48.539051085 +0000 UTC m=+2105.281761680" observedRunningTime="2025-12-03 22:16:48.97188393 +0000 UTC m=+2105.714594535" watchObservedRunningTime="2025-12-03 22:16:48.98273581 +0000 UTC m=+2105.725446415" Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.072868 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-kp478"] Dec 03 22:16:49 crc kubenswrapper[4715]: W1203 22:16:49.079738 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60edd1a1_9b8a_4d7d_977b_11c834922ca1.slice/crio-1727a2236c12e8c6659f466899c57d66d1fd60c88c5d475b538b4ca920f00302 WatchSource:0}: Error finding container 1727a2236c12e8c6659f466899c57d66d1fd60c88c5d475b538b4ca920f00302: Status 404 returned error can't find the container with id 1727a2236c12e8c6659f466899c57d66d1fd60c88c5d475b538b4ca920f00302 Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.084333 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.198334 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74cd84ccf4-k9bcn"] Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.223883 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-656478594d-5tbrm"] Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.305691 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:16:49 crc kubenswrapper[4715]: W1203 22:16:49.608481 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c7cfacf_a283_45d3_982d_b28e62a9146b.slice/crio-ef54c16ba114a2e67230b8162fcf2dc66a1fefaf952375e3cd5ff5ac0af9405a WatchSource:0}: Error finding container ef54c16ba114a2e67230b8162fcf2dc66a1fefaf952375e3cd5ff5ac0af9405a: Status 404 returned error can't find the container with id ef54c16ba114a2e67230b8162fcf2dc66a1fefaf952375e3cd5ff5ac0af9405a Dec 03 22:16:49 crc kubenswrapper[4715]: W1203 22:16:49.615388 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae09bc5f_d3e1_48c3_8145_611fd9827870.slice/crio-2e99021e5e3860269e29c5cca98f3ef145efad242585dad019ddf055dff63714 WatchSource:0}: Error finding container 2e99021e5e3860269e29c5cca98f3ef145efad242585dad019ddf055dff63714: Status 404 returned error can't find the container with id 2e99021e5e3860269e29c5cca98f3ef145efad242585dad019ddf055dff63714 Dec 03 22:16:49 crc kubenswrapper[4715]: W1203 22:16:49.625755 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod279fe539_9749_4528_b219_df6323816ea7.slice/crio-176af77e30c5143c1fc7b70bcecfab63b2a76e856708458ea11029e0d24f963e WatchSource:0}: Error finding container 176af77e30c5143c1fc7b70bcecfab63b2a76e856708458ea11029e0d24f963e: Status 404 returned error can't find the container with id 176af77e30c5143c1fc7b70bcecfab63b2a76e856708458ea11029e0d24f963e Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.647327 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="391549ec-5cda-4f77-917c-0b485e1d57bf" path="/var/lib/kubelet/pods/391549ec-5cda-4f77-917c-0b485e1d57bf/volumes" Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.648666 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da4847d8-02fb-41f4-8a71-0cf1eee4f55b" path="/var/lib/kubelet/pods/da4847d8-02fb-41f4-8a71-0cf1eee4f55b/volumes" Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.969450 4715 generic.go:334] "Generic (PLEG): container finished" podID="6ded23d0-fc8b-4ed7-8358-573da1f24919" containerID="371aa80cbef5669afbd1d3dc6067b637a7c1618fca9e34f0e35de061c608b7c8" exitCode=0 Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.969554 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-blfqt" event={"ID":"6ded23d0-fc8b-4ed7-8358-573da1f24919","Type":"ContainerDied","Data":"371aa80cbef5669afbd1d3dc6067b637a7c1618fca9e34f0e35de061c608b7c8"} Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.975062 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656478594d-5tbrm" event={"ID":"6c7cfacf-a283-45d3-982d-b28e62a9146b","Type":"ContainerStarted","Data":"ef54c16ba114a2e67230b8162fcf2dc66a1fefaf952375e3cd5ff5ac0af9405a"} Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.981811 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cd84ccf4-k9bcn" event={"ID":"ae09bc5f-d3e1-48c3-8145-611fd9827870","Type":"ContainerStarted","Data":"2e99021e5e3860269e29c5cca98f3ef145efad242585dad019ddf055dff63714"} Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.984072 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kp478" event={"ID":"60edd1a1-9b8a-4d7d-977b-11c834922ca1","Type":"ContainerStarted","Data":"ad4f3a024bad21cc3c9bcda15545d8af611a2f98e124ad9fcb76eb260681987f"} Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.984141 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kp478" event={"ID":"60edd1a1-9b8a-4d7d-977b-11c834922ca1","Type":"ContainerStarted","Data":"1727a2236c12e8c6659f466899c57d66d1fd60c88c5d475b538b4ca920f00302"} Dec 03 22:16:49 crc kubenswrapper[4715]: I1203 22:16:49.988821 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"279fe539-9749-4528-b219-df6323816ea7","Type":"ContainerStarted","Data":"176af77e30c5143c1fc7b70bcecfab63b2a76e856708458ea11029e0d24f963e"} Dec 03 22:16:50 crc kubenswrapper[4715]: I1203 22:16:50.010107 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-kp478" podStartSLOduration=18.010090593 podStartE2EDuration="18.010090593s" podCreationTimestamp="2025-12-03 22:16:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:16:50.004407381 +0000 UTC m=+2106.747118016" watchObservedRunningTime="2025-12-03 22:16:50.010090593 +0000 UTC m=+2106.752801188" Dec 03 22:16:50 crc kubenswrapper[4715]: I1203 22:16:50.425653 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.001301 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6ngmv" event={"ID":"2eba90e4-4991-4260-9e8e-2a3a5412b264","Type":"ContainerStarted","Data":"8b4284641a08341e55d09f63043ef7d4f7d7d31499b72a5302b241474ad97d4a"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.003546 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cd84ccf4-k9bcn" event={"ID":"ae09bc5f-d3e1-48c3-8145-611fd9827870","Type":"ContainerStarted","Data":"4979bdb653c8db320223f4fb81a9d23bee0c19d8f6be47165c7af4d53c639d1f"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.003575 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cd84ccf4-k9bcn" event={"ID":"ae09bc5f-d3e1-48c3-8145-611fd9827870","Type":"ContainerStarted","Data":"706dc959898b0c6a666acb6b020a24e81832e5be7099c366f91458474d2b0618"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.005854 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"279fe539-9749-4528-b219-df6323816ea7","Type":"ContainerStarted","Data":"f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.011737 4715 generic.go:334] "Generic (PLEG): container finished" podID="d3e2fa00-220e-4811-8b41-644d96b70a46" containerID="3fa789e13ceb6805ffdd0b41900fc86eed3857e22eaf5896774011c906deb354" exitCode=0 Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.011779 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-d8zcz" event={"ID":"d3e2fa00-220e-4811-8b41-644d96b70a46","Type":"ContainerDied","Data":"3fa789e13ceb6805ffdd0b41900fc86eed3857e22eaf5896774011c906deb354"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.016753 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-6ngmv" podStartSLOduration=3.431471846 podStartE2EDuration="43.016738902s" podCreationTimestamp="2025-12-03 22:16:08 +0000 UTC" firstStartedPulling="2025-12-03 22:16:10.319310846 +0000 UTC m=+2067.062021441" lastFinishedPulling="2025-12-03 22:16:49.904577902 +0000 UTC m=+2106.647288497" observedRunningTime="2025-12-03 22:16:51.012592811 +0000 UTC m=+2107.755303406" watchObservedRunningTime="2025-12-03 22:16:51.016738902 +0000 UTC m=+2107.759449497" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.023578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656478594d-5tbrm" event={"ID":"6c7cfacf-a283-45d3-982d-b28e62a9146b","Type":"ContainerStarted","Data":"289fcdafec040cd54e580cc2c8d312a98fcfda1df2934208ae6a531e17bf73b3"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.023607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-656478594d-5tbrm" event={"ID":"6c7cfacf-a283-45d3-982d-b28e62a9146b","Type":"ContainerStarted","Data":"486d23cbb8c56977a3b10fc264235f804ec6e32675990d96a991b74e52bd5c91"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.031551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6143cec8-8a39-443c-b850-109e24ac1200","Type":"ContainerStarted","Data":"3c2056cd8e993c9e3b9a3c5c5e37e0849818a26f5d0eceabf1165ac94696fa2f"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.031592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6143cec8-8a39-443c-b850-109e24ac1200","Type":"ContainerStarted","Data":"9081f3e7e372b94935f2f4a428701e17b45edccd1700f7600cfcfb329b7e0215"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.038554 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72b03167-1df1-47c3-94c7-e0f13b65f972","Type":"ContainerStarted","Data":"0239ba4c5711c4500c470cdc0c11f24c683374f0d887c7aa118733dae3186d74"} Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.048784 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74cd84ccf4-k9bcn" podStartSLOduration=33.578144114 podStartE2EDuration="34.048768459s" podCreationTimestamp="2025-12-03 22:16:17 +0000 UTC" firstStartedPulling="2025-12-03 22:16:49.617571547 +0000 UTC m=+2106.360282182" lastFinishedPulling="2025-12-03 22:16:50.088195932 +0000 UTC m=+2106.830906527" observedRunningTime="2025-12-03 22:16:51.046532978 +0000 UTC m=+2107.789243563" watchObservedRunningTime="2025-12-03 22:16:51.048768459 +0000 UTC m=+2107.791479054" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.315636 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.328492 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-656478594d-5tbrm" podStartSLOduration=32.730519939 podStartE2EDuration="33.328477229s" podCreationTimestamp="2025-12-03 22:16:18 +0000 UTC" firstStartedPulling="2025-12-03 22:16:49.614457703 +0000 UTC m=+2106.357168338" lastFinishedPulling="2025-12-03 22:16:50.212415013 +0000 UTC m=+2106.955125628" observedRunningTime="2025-12-03 22:16:51.07764998 +0000 UTC m=+2107.820360575" watchObservedRunningTime="2025-12-03 22:16:51.328477229 +0000 UTC m=+2108.071187824" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.395285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-combined-ca-bundle\") pod \"6ded23d0-fc8b-4ed7-8358-573da1f24919\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.395452 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8kmz\" (UniqueName: \"kubernetes.io/projected/6ded23d0-fc8b-4ed7-8358-573da1f24919-kube-api-access-c8kmz\") pod \"6ded23d0-fc8b-4ed7-8358-573da1f24919\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.395605 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-config\") pod \"6ded23d0-fc8b-4ed7-8358-573da1f24919\" (UID: \"6ded23d0-fc8b-4ed7-8358-573da1f24919\") " Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.401654 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ded23d0-fc8b-4ed7-8358-573da1f24919-kube-api-access-c8kmz" (OuterVolumeSpecName: "kube-api-access-c8kmz") pod "6ded23d0-fc8b-4ed7-8358-573da1f24919" (UID: "6ded23d0-fc8b-4ed7-8358-573da1f24919"). InnerVolumeSpecName "kube-api-access-c8kmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.421809 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-config" (OuterVolumeSpecName: "config") pod "6ded23d0-fc8b-4ed7-8358-573da1f24919" (UID: "6ded23d0-fc8b-4ed7-8358-573da1f24919"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.425777 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6ded23d0-fc8b-4ed7-8358-573da1f24919" (UID: "6ded23d0-fc8b-4ed7-8358-573da1f24919"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.497483 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.497533 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6ded23d0-fc8b-4ed7-8358-573da1f24919-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:51 crc kubenswrapper[4715]: I1203 22:16:51.497545 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8kmz\" (UniqueName: \"kubernetes.io/projected/6ded23d0-fc8b-4ed7-8358-573da1f24919-kube-api-access-c8kmz\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.051518 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"279fe539-9749-4528-b219-df6323816ea7","Type":"ContainerStarted","Data":"a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f"} Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.055010 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-blfqt" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.055090 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-blfqt" event={"ID":"6ded23d0-fc8b-4ed7-8358-573da1f24919","Type":"ContainerDied","Data":"de781bce412099648733b54ebfd61580f0ee94df7e9a0af6e62cdaaf1b7d4a11"} Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.055451 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de781bce412099648733b54ebfd61580f0ee94df7e9a0af6e62cdaaf1b7d4a11" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.059144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6143cec8-8a39-443c-b850-109e24ac1200","Type":"ContainerStarted","Data":"dba4177ff9d25999c348f58f9bc7a868140faaff34981749ae6ecb131064a243"} Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.065336 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.065371 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.065383 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.065394 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.084114 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.084094614 podStartE2EDuration="5.084094614s" podCreationTimestamp="2025-12-03 22:16:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:16:52.074639161 +0000 UTC m=+2108.817349846" watchObservedRunningTime="2025-12-03 22:16:52.084094614 +0000 UTC m=+2108.826805209" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.112793 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=31.112769851 podStartE2EDuration="31.112769851s" podCreationTimestamp="2025-12-03 22:16:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:16:52.10601267 +0000 UTC m=+2108.848723255" watchObservedRunningTime="2025-12-03 22:16:52.112769851 +0000 UTC m=+2108.855480456" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.137993 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.147602 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.263430 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-7fj29"] Dec 03 22:16:52 crc kubenswrapper[4715]: E1203 22:16:52.263937 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ded23d0-fc8b-4ed7-8358-573da1f24919" containerName="neutron-db-sync" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.263961 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ded23d0-fc8b-4ed7-8358-573da1f24919" containerName="neutron-db-sync" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.264167 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ded23d0-fc8b-4ed7-8358-573da1f24919" containerName="neutron-db-sync" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.265339 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.283019 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-7fj29"] Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.425338 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-svc\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.425379 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-config\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.425553 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.425586 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.425694 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.425921 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-644dm\" (UniqueName: \"kubernetes.io/projected/b197785c-154d-4e76-a57b-8310513175ba-kube-api-access-644dm\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.471687 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8c9bc67fb-47rvm"] Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.475868 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.478664 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-n9b5g" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.478844 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.478990 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.479124 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.491612 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8c9bc67fb-47rvm"] Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.527455 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-svc\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.527516 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-config\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.527563 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.527586 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.527611 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.527669 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-644dm\" (UniqueName: \"kubernetes.io/projected/b197785c-154d-4e76-a57b-8310513175ba-kube-api-access-644dm\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.528748 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.529094 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.529270 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.529852 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-svc\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.533039 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-config\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.553222 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-644dm\" (UniqueName: \"kubernetes.io/projected/b197785c-154d-4e76-a57b-8310513175ba-kube-api-access-644dm\") pod \"dnsmasq-dns-6b7b667979-7fj29\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.609960 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.628433 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-ovndb-tls-certs\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.628528 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-config\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.628552 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-httpd-config\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.628609 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-combined-ca-bundle\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.628633 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4ph5\" (UniqueName: \"kubernetes.io/projected/de1bec8a-6677-4be3-8722-5ec1bb336d6f-kube-api-access-z4ph5\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.629770 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.730291 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e2fa00-220e-4811-8b41-644d96b70a46-logs\") pod \"d3e2fa00-220e-4811-8b41-644d96b70a46\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.730421 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgtrr\" (UniqueName: \"kubernetes.io/projected/d3e2fa00-220e-4811-8b41-644d96b70a46-kube-api-access-cgtrr\") pod \"d3e2fa00-220e-4811-8b41-644d96b70a46\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.730449 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-scripts\") pod \"d3e2fa00-220e-4811-8b41-644d96b70a46\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.730479 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-combined-ca-bundle\") pod \"d3e2fa00-220e-4811-8b41-644d96b70a46\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.730649 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-config-data\") pod \"d3e2fa00-220e-4811-8b41-644d96b70a46\" (UID: \"d3e2fa00-220e-4811-8b41-644d96b70a46\") " Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.730932 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4ph5\" (UniqueName: \"kubernetes.io/projected/de1bec8a-6677-4be3-8722-5ec1bb336d6f-kube-api-access-z4ph5\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.730995 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-ovndb-tls-certs\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.731079 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-config\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.731114 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-httpd-config\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.731213 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-combined-ca-bundle\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.732061 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3e2fa00-220e-4811-8b41-644d96b70a46-logs" (OuterVolumeSpecName: "logs") pod "d3e2fa00-220e-4811-8b41-644d96b70a46" (UID: "d3e2fa00-220e-4811-8b41-644d96b70a46"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.734663 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3e2fa00-220e-4811-8b41-644d96b70a46-kube-api-access-cgtrr" (OuterVolumeSpecName: "kube-api-access-cgtrr") pod "d3e2fa00-220e-4811-8b41-644d96b70a46" (UID: "d3e2fa00-220e-4811-8b41-644d96b70a46"). InnerVolumeSpecName "kube-api-access-cgtrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.734883 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-combined-ca-bundle\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.738929 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-httpd-config\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.740684 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-config\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.743646 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-scripts" (OuterVolumeSpecName: "scripts") pod "d3e2fa00-220e-4811-8b41-644d96b70a46" (UID: "d3e2fa00-220e-4811-8b41-644d96b70a46"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.753948 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4ph5\" (UniqueName: \"kubernetes.io/projected/de1bec8a-6677-4be3-8722-5ec1bb336d6f-kube-api-access-z4ph5\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.761135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-ovndb-tls-certs\") pod \"neutron-8c9bc67fb-47rvm\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.778683 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3e2fa00-220e-4811-8b41-644d96b70a46" (UID: "d3e2fa00-220e-4811-8b41-644d96b70a46"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.782849 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-config-data" (OuterVolumeSpecName: "config-data") pod "d3e2fa00-220e-4811-8b41-644d96b70a46" (UID: "d3e2fa00-220e-4811-8b41-644d96b70a46"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.832782 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.832823 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3e2fa00-220e-4811-8b41-644d96b70a46-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.832833 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgtrr\" (UniqueName: \"kubernetes.io/projected/d3e2fa00-220e-4811-8b41-644d96b70a46-kube-api-access-cgtrr\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.832844 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.832851 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3e2fa00-220e-4811-8b41-644d96b70a46-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.889012 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-7fj29"] Dec 03 22:16:52 crc kubenswrapper[4715]: W1203 22:16:52.913726 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb197785c_154d_4e76_a57b_8310513175ba.slice/crio-4c436b16395a56f7796a58be73abb04fd8548aa0d24d53acbca95f8f0ca635d9 WatchSource:0}: Error finding container 4c436b16395a56f7796a58be73abb04fd8548aa0d24d53acbca95f8f0ca635d9: Status 404 returned error can't find the container with id 4c436b16395a56f7796a58be73abb04fd8548aa0d24d53acbca95f8f0ca635d9 Dec 03 22:16:52 crc kubenswrapper[4715]: I1203 22:16:52.926904 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.119551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-d8zcz" event={"ID":"d3e2fa00-220e-4811-8b41-644d96b70a46","Type":"ContainerDied","Data":"1e1b216fe46d6b2ff507fda8255f6644a7f51d023f7bfd4307c0feb95ba11f59"} Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.119611 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e1b216fe46d6b2ff507fda8255f6644a7f51d023f7bfd4307c0feb95ba11f59" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.119693 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-d8zcz" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.125228 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" event={"ID":"b197785c-154d-4e76-a57b-8310513175ba","Type":"ContainerStarted","Data":"4c436b16395a56f7796a58be73abb04fd8548aa0d24d53acbca95f8f0ca635d9"} Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.186647 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8694d8458b-2x672"] Dec 03 22:16:53 crc kubenswrapper[4715]: E1203 22:16:53.187387 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3e2fa00-220e-4811-8b41-644d96b70a46" containerName="placement-db-sync" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.187412 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3e2fa00-220e-4811-8b41-644d96b70a46" containerName="placement-db-sync" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.187718 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3e2fa00-220e-4811-8b41-644d96b70a46" containerName="placement-db-sync" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.188920 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.191395 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.191560 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.191680 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.191784 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-rhz28" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.191888 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.193243 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8694d8458b-2x672"] Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.345655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/661b1086-fc52-475a-a449-c116baa618a2-logs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.345703 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-config-data\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.345737 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-internal-tls-certs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.345771 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-combined-ca-bundle\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.345799 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxvx9\" (UniqueName: \"kubernetes.io/projected/661b1086-fc52-475a-a449-c116baa618a2-kube-api-access-gxvx9\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.345868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-scripts\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.345900 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-public-tls-certs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.447973 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-combined-ca-bundle\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.448040 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxvx9\" (UniqueName: \"kubernetes.io/projected/661b1086-fc52-475a-a449-c116baa618a2-kube-api-access-gxvx9\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.448085 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-scripts\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.448114 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-public-tls-certs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.448246 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/661b1086-fc52-475a-a449-c116baa618a2-logs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.448277 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-config-data\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.448321 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-internal-tls-certs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.448844 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/661b1086-fc52-475a-a449-c116baa618a2-logs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.456197 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-combined-ca-bundle\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.457076 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-public-tls-certs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.458467 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-config-data\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.458571 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-scripts\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.461347 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/661b1086-fc52-475a-a449-c116baa618a2-internal-tls-certs\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.464178 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxvx9\" (UniqueName: \"kubernetes.io/projected/661b1086-fc52-475a-a449-c116baa618a2-kube-api-access-gxvx9\") pod \"placement-8694d8458b-2x672\" (UID: \"661b1086-fc52-475a-a449-c116baa618a2\") " pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.562415 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8694d8458b-2x672" Dec 03 22:16:53 crc kubenswrapper[4715]: I1203 22:16:53.597936 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8c9bc67fb-47rvm"] Dec 03 22:16:53 crc kubenswrapper[4715]: W1203 22:16:53.614430 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde1bec8a_6677_4be3_8722_5ec1bb336d6f.slice/crio-b90af023126aeb6fe6fecba141c3ec5af87623d45986777aa4c6cd3a134ed90b WatchSource:0}: Error finding container b90af023126aeb6fe6fecba141c3ec5af87623d45986777aa4c6cd3a134ed90b: Status 404 returned error can't find the container with id b90af023126aeb6fe6fecba141c3ec5af87623d45986777aa4c6cd3a134ed90b Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.068668 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8694d8458b-2x672"] Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.138032 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c9bc67fb-47rvm" event={"ID":"de1bec8a-6677-4be3-8722-5ec1bb336d6f","Type":"ContainerStarted","Data":"b90af023126aeb6fe6fecba141c3ec5af87623d45986777aa4c6cd3a134ed90b"} Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.145240 4715 generic.go:334] "Generic (PLEG): container finished" podID="b197785c-154d-4e76-a57b-8310513175ba" containerID="9179ccf2edf7b5df92e12172e0eaddac8184a42ab9f884bd8d60549d991811c9" exitCode=0 Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.145444 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" event={"ID":"b197785c-154d-4e76-a57b-8310513175ba","Type":"ContainerDied","Data":"9179ccf2edf7b5df92e12172e0eaddac8184a42ab9f884bd8d60549d991811c9"} Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.757112 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-666d579d9-hnzld"] Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.758905 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.766066 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.766479 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.784089 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-666d579d9-hnzld"] Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.878144 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-ovndb-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.878234 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-config\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.878258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-internal-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.878327 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-httpd-config\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.878392 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xszw\" (UniqueName: \"kubernetes.io/projected/58badebd-4a34-4430-96ee-706deee12007-kube-api-access-2xszw\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.878456 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-public-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.878527 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-combined-ca-bundle\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.980239 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-combined-ca-bundle\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.980341 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-ovndb-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.980372 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-config\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.980393 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-internal-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.980410 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-httpd-config\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.980438 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xszw\" (UniqueName: \"kubernetes.io/projected/58badebd-4a34-4430-96ee-706deee12007-kube-api-access-2xszw\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.980484 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-public-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.986929 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-public-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.987258 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-combined-ca-bundle\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.987331 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-internal-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.988150 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-ovndb-tls-certs\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.988342 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-config\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.988416 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/58badebd-4a34-4430-96ee-706deee12007-httpd-config\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:54 crc kubenswrapper[4715]: I1203 22:16:54.997104 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xszw\" (UniqueName: \"kubernetes.io/projected/58badebd-4a34-4430-96ee-706deee12007-kube-api-access-2xszw\") pod \"neutron-666d579d9-hnzld\" (UID: \"58badebd-4a34-4430-96ee-706deee12007\") " pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:55 crc kubenswrapper[4715]: I1203 22:16:55.085896 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:16:55 crc kubenswrapper[4715]: I1203 22:16:55.156165 4715 generic.go:334] "Generic (PLEG): container finished" podID="2eba90e4-4991-4260-9e8e-2a3a5412b264" containerID="8b4284641a08341e55d09f63043ef7d4f7d7d31499b72a5302b241474ad97d4a" exitCode=0 Dec 03 22:16:55 crc kubenswrapper[4715]: I1203 22:16:55.156232 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6ngmv" event={"ID":"2eba90e4-4991-4260-9e8e-2a3a5412b264","Type":"ContainerDied","Data":"8b4284641a08341e55d09f63043ef7d4f7d7d31499b72a5302b241474ad97d4a"} Dec 03 22:16:55 crc kubenswrapper[4715]: I1203 22:16:55.158074 4715 generic.go:334] "Generic (PLEG): container finished" podID="60edd1a1-9b8a-4d7d-977b-11c834922ca1" containerID="ad4f3a024bad21cc3c9bcda15545d8af611a2f98e124ad9fcb76eb260681987f" exitCode=0 Dec 03 22:16:55 crc kubenswrapper[4715]: I1203 22:16:55.158102 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kp478" event={"ID":"60edd1a1-9b8a-4d7d-977b-11c834922ca1","Type":"ContainerDied","Data":"ad4f3a024bad21cc3c9bcda15545d8af611a2f98e124ad9fcb76eb260681987f"} Dec 03 22:16:55 crc kubenswrapper[4715]: W1203 22:16:55.566406 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod661b1086_fc52_475a_a449_c116baa618a2.slice/crio-b8d30aff8716a596fdfbbfa736d06e3c4c4a6cf2332a0ae75c304893ccd9232f WatchSource:0}: Error finding container b8d30aff8716a596fdfbbfa736d06e3c4c4a6cf2332a0ae75c304893ccd9232f: Status 404 returned error can't find the container with id b8d30aff8716a596fdfbbfa736d06e3c4c4a6cf2332a0ae75c304893ccd9232f Dec 03 22:16:56 crc kubenswrapper[4715]: I1203 22:16:56.169279 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8694d8458b-2x672" event={"ID":"661b1086-fc52-475a-a449-c116baa618a2","Type":"ContainerStarted","Data":"b8d30aff8716a596fdfbbfa736d06e3c4c4a6cf2332a0ae75c304893ccd9232f"} Dec 03 22:16:58 crc kubenswrapper[4715]: I1203 22:16:58.313927 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:58 crc kubenswrapper[4715]: I1203 22:16:58.314481 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:16:58 crc kubenswrapper[4715]: I1203 22:16:58.356978 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:58 crc kubenswrapper[4715]: I1203 22:16:58.357453 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:58 crc kubenswrapper[4715]: I1203 22:16:58.410701 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:58 crc kubenswrapper[4715]: I1203 22:16:58.422766 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:58 crc kubenswrapper[4715]: I1203 22:16:58.511941 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:58 crc kubenswrapper[4715]: I1203 22:16:58.511989 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:16:59 crc kubenswrapper[4715]: I1203 22:16:59.205703 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 22:16:59 crc kubenswrapper[4715]: I1203 22:16:59.207420 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.933681 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kp478" Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.973678 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.988769 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-scripts\") pod \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.988856 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-config-data\") pod \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.988885 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-db-sync-config-data\") pod \"2eba90e4-4991-4260-9e8e-2a3a5412b264\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.988949 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gjjcm\" (UniqueName: \"kubernetes.io/projected/60edd1a1-9b8a-4d7d-977b-11c834922ca1-kube-api-access-gjjcm\") pod \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.989000 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsskp\" (UniqueName: \"kubernetes.io/projected/2eba90e4-4991-4260-9e8e-2a3a5412b264-kube-api-access-dsskp\") pod \"2eba90e4-4991-4260-9e8e-2a3a5412b264\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.989100 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-combined-ca-bundle\") pod \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.989197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-fernet-keys\") pod \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.989246 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-combined-ca-bundle\") pod \"2eba90e4-4991-4260-9e8e-2a3a5412b264\" (UID: \"2eba90e4-4991-4260-9e8e-2a3a5412b264\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.989281 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-credential-keys\") pod \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\" (UID: \"60edd1a1-9b8a-4d7d-977b-11c834922ca1\") " Dec 03 22:17:00 crc kubenswrapper[4715]: I1203 22:17:00.997716 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60edd1a1-9b8a-4d7d-977b-11c834922ca1-kube-api-access-gjjcm" (OuterVolumeSpecName: "kube-api-access-gjjcm") pod "60edd1a1-9b8a-4d7d-977b-11c834922ca1" (UID: "60edd1a1-9b8a-4d7d-977b-11c834922ca1"). InnerVolumeSpecName "kube-api-access-gjjcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.002196 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2eba90e4-4991-4260-9e8e-2a3a5412b264" (UID: "2eba90e4-4991-4260-9e8e-2a3a5412b264"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.002217 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-scripts" (OuterVolumeSpecName: "scripts") pod "60edd1a1-9b8a-4d7d-977b-11c834922ca1" (UID: "60edd1a1-9b8a-4d7d-977b-11c834922ca1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.002382 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "60edd1a1-9b8a-4d7d-977b-11c834922ca1" (UID: "60edd1a1-9b8a-4d7d-977b-11c834922ca1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.002599 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "60edd1a1-9b8a-4d7d-977b-11c834922ca1" (UID: "60edd1a1-9b8a-4d7d-977b-11c834922ca1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.004677 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2eba90e4-4991-4260-9e8e-2a3a5412b264-kube-api-access-dsskp" (OuterVolumeSpecName: "kube-api-access-dsskp") pod "2eba90e4-4991-4260-9e8e-2a3a5412b264" (UID: "2eba90e4-4991-4260-9e8e-2a3a5412b264"). InnerVolumeSpecName "kube-api-access-dsskp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.043905 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-config-data" (OuterVolumeSpecName: "config-data") pod "60edd1a1-9b8a-4d7d-977b-11c834922ca1" (UID: "60edd1a1-9b8a-4d7d-977b-11c834922ca1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.105400 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gjjcm\" (UniqueName: \"kubernetes.io/projected/60edd1a1-9b8a-4d7d-977b-11c834922ca1-kube-api-access-gjjcm\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.105435 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsskp\" (UniqueName: \"kubernetes.io/projected/2eba90e4-4991-4260-9e8e-2a3a5412b264-kube-api-access-dsskp\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.105446 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.105457 4715 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.105466 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.105474 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.105483 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.115354 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2eba90e4-4991-4260-9e8e-2a3a5412b264" (UID: "2eba90e4-4991-4260-9e8e-2a3a5412b264"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.117942 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60edd1a1-9b8a-4d7d-977b-11c834922ca1" (UID: "60edd1a1-9b8a-4d7d-977b-11c834922ca1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.206782 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60edd1a1-9b8a-4d7d-977b-11c834922ca1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.206807 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2eba90e4-4991-4260-9e8e-2a3a5412b264-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.222720 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c9bc67fb-47rvm" event={"ID":"de1bec8a-6677-4be3-8722-5ec1bb336d6f","Type":"ContainerStarted","Data":"f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e"} Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.232912 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72b03167-1df1-47c3-94c7-e0f13b65f972","Type":"ContainerStarted","Data":"126f2d956398730e33ce6940fe54f6d10ae8f579570da1802dd5231d55ff29a9"} Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.236291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8694d8458b-2x672" event={"ID":"661b1086-fc52-475a-a449-c116baa618a2","Type":"ContainerStarted","Data":"adef1e7fd6f3138d560333583f98d2a5f7f2f3da39542b35b48e0b7d1e44bde8"} Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.236901 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-666d579d9-hnzld"] Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.248534 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" event={"ID":"b197785c-154d-4e76-a57b-8310513175ba","Type":"ContainerStarted","Data":"6a895988eed56b92a122486af19bc098702fedc2bc115239bf4827819508a394"} Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.249602 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.253340 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-6ngmv" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.253494 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-6ngmv" event={"ID":"2eba90e4-4991-4260-9e8e-2a3a5412b264","Type":"ContainerDied","Data":"4e7d91578a5e8d5d9ed60a8225035599b07e3a00d2b8f3477a7cf87554210224"} Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.253535 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e7d91578a5e8d5d9ed60a8225035599b07e3a00d2b8f3477a7cf87554210224" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.255077 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.255104 4715 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.255993 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-kp478" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.257380 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-kp478" event={"ID":"60edd1a1-9b8a-4d7d-977b-11c834922ca1","Type":"ContainerDied","Data":"1727a2236c12e8c6659f466899c57d66d1fd60c88c5d475b538b4ca920f00302"} Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.257409 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1727a2236c12e8c6659f466899c57d66d1fd60c88c5d475b538b4ca920f00302" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.293304 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" podStartSLOduration=9.293282589 podStartE2EDuration="9.293282589s" podCreationTimestamp="2025-12-03 22:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:01.284849693 +0000 UTC m=+2118.027560298" watchObservedRunningTime="2025-12-03 22:17:01.293282589 +0000 UTC m=+2118.035993184" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.686659 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 22:17:01 crc kubenswrapper[4715]: I1203 22:17:01.754464 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.181185 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-76bcbf7d4d-l7qmc"] Dec 03 22:17:02 crc kubenswrapper[4715]: E1203 22:17:02.181536 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60edd1a1-9b8a-4d7d-977b-11c834922ca1" containerName="keystone-bootstrap" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.181547 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="60edd1a1-9b8a-4d7d-977b-11c834922ca1" containerName="keystone-bootstrap" Dec 03 22:17:02 crc kubenswrapper[4715]: E1203 22:17:02.181565 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2eba90e4-4991-4260-9e8e-2a3a5412b264" containerName="barbican-db-sync" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.181571 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2eba90e4-4991-4260-9e8e-2a3a5412b264" containerName="barbican-db-sync" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.181720 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2eba90e4-4991-4260-9e8e-2a3a5412b264" containerName="barbican-db-sync" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.181737 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="60edd1a1-9b8a-4d7d-977b-11c834922ca1" containerName="keystone-bootstrap" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.182277 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.185166 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.185246 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.185170 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.186879 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.186940 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-9bcdn" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.186950 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.196912 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-76bcbf7d4d-l7qmc"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.308815 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p4mjk" event={"ID":"5ec32c20-9117-4eca-9b73-a7f8b84677da","Type":"ContainerStarted","Data":"16fd66c726542fba27a9cf3c4da98d61f54d7099995437d0013b10cd48184d2a"} Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.319736 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-78449cf9c5-v4vnx"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.321272 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.329846 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mwckq" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.330041 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.330167 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.337390 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-9cc86b6b-6xlfj"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.339079 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.343923 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.346931 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c9bc67fb-47rvm" event={"ID":"de1bec8a-6677-4be3-8722-5ec1bb336d6f","Type":"ContainerStarted","Data":"ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b"} Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.347150 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356259 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-scripts\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356349 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5k58\" (UniqueName: \"kubernetes.io/projected/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-kube-api-access-p5k58\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356349 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-9cc86b6b-6xlfj"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356379 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-fernet-keys\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356426 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-credential-keys\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356468 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-public-tls-certs\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356490 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-internal-tls-certs\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356520 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-config-data\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.356538 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-combined-ca-bundle\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.358351 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8694d8458b-2x672" event={"ID":"661b1086-fc52-475a-a449-c116baa618a2","Type":"ContainerStarted","Data":"3a86fdaa0f7f723dc5c3669e23f89cb30f9393d708e99dd96b545b7bfb2c992d"} Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.358385 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8694d8458b-2x672" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.358407 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-8694d8458b-2x672" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.382886 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78449cf9c5-v4vnx"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.391002 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666d579d9-hnzld" event={"ID":"58badebd-4a34-4430-96ee-706deee12007","Type":"ContainerStarted","Data":"fb0a78f345cb7d4f8a5c91abfc5c6bd40dd0d2fdfcc5ed31adce73448c3a8ca6"} Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.391039 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666d579d9-hnzld" event={"ID":"58badebd-4a34-4430-96ee-706deee12007","Type":"ContainerStarted","Data":"d4a14e379aeb0d4698b4f1452a24e2ba160b5285fa03b04492907f5f019b25a0"} Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.391054 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.391065 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-666d579d9-hnzld" event={"ID":"58badebd-4a34-4430-96ee-706deee12007","Type":"ContainerStarted","Data":"5f66d8c86b493085d4b65e0fde0ee4bf6ce4addfd35a6c24c64e0b195a170b6b"} Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.407927 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-p4mjk" podStartSLOduration=3.579695041 podStartE2EDuration="54.407905895s" podCreationTimestamp="2025-12-03 22:16:08 +0000 UTC" firstStartedPulling="2025-12-03 22:16:10.319225184 +0000 UTC m=+2067.061935779" lastFinishedPulling="2025-12-03 22:17:01.147436038 +0000 UTC m=+2117.890146633" observedRunningTime="2025-12-03 22:17:02.330373501 +0000 UTC m=+2119.073084096" watchObservedRunningTime="2025-12-03 22:17:02.407905895 +0000 UTC m=+2119.150616490" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458125 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-fernet-keys\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458180 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-combined-ca-bundle\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458247 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-credential-keys\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458287 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvvwt\" (UniqueName: \"kubernetes.io/projected/4382c9dd-8a2d-4634-b9de-0796a994d491-kube-api-access-qvvwt\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458335 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thtjg\" (UniqueName: \"kubernetes.io/projected/0e740246-a3e9-4692-8888-e12768f0573e-kube-api-access-thtjg\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458381 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-config-data\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458406 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-config-data-custom\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458431 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-public-tls-certs\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458455 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-combined-ca-bundle\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458493 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-internal-tls-certs\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458578 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-config-data\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-combined-ca-bundle\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458656 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4382c9dd-8a2d-4634-b9de-0796a994d491-logs\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458696 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-config-data\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458733 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-scripts\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458768 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e740246-a3e9-4692-8888-e12768f0573e-logs\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-config-data-custom\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.458934 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5k58\" (UniqueName: \"kubernetes.io/projected/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-kube-api-access-p5k58\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.478603 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-credential-keys\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.484490 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-public-tls-certs\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.485981 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-config-data\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.501481 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-7fj29"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.503880 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-scripts\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.514674 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-internal-tls-certs\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.516260 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-combined-ca-bundle\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.533260 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-fernet-keys\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.533741 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5k58\" (UniqueName: \"kubernetes.io/projected/954b69af-5e1b-4eec-a783-1cbe28e7c0ca-kube-api-access-p5k58\") pod \"keystone-76bcbf7d4d-l7qmc\" (UID: \"954b69af-5e1b-4eec-a783-1cbe28e7c0ca\") " pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.550934 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7f9c554bf8-zrldg"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.554655 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.559120 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-config-data\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560132 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-config-data-custom\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560188 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-combined-ca-bundle\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560245 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4382c9dd-8a2d-4634-b9de-0796a994d491-logs\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-config-data\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560327 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e740246-a3e9-4692-8888-e12768f0573e-logs\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560416 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-config-data-custom\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560478 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-combined-ca-bundle\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560533 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvvwt\" (UniqueName: \"kubernetes.io/projected/4382c9dd-8a2d-4634-b9de-0796a994d491-kube-api-access-qvvwt\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.560565 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thtjg\" (UniqueName: \"kubernetes.io/projected/0e740246-a3e9-4692-8888-e12768f0573e-kube-api-access-thtjg\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.562939 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-8c9bc67fb-47rvm" podStartSLOduration=10.56292973 podStartE2EDuration="10.56292973s" podCreationTimestamp="2025-12-03 22:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:02.43840058 +0000 UTC m=+2119.181111175" watchObservedRunningTime="2025-12-03 22:17:02.56292973 +0000 UTC m=+2119.305640325" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.564192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4382c9dd-8a2d-4634-b9de-0796a994d491-logs\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.564396 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e740246-a3e9-4692-8888-e12768f0573e-logs\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.569081 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-combined-ca-bundle\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.575134 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-config-data\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.578564 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f9c554bf8-zrldg"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.581124 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-config-data-custom\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.581680 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-config-data-custom\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.584406 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4382c9dd-8a2d-4634-b9de-0796a994d491-combined-ca-bundle\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.585603 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0e740246-a3e9-4692-8888-e12768f0573e-config-data\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.587211 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-srvlb"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.587926 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-8694d8458b-2x672" podStartSLOduration=9.587911428 podStartE2EDuration="9.587911428s" podCreationTimestamp="2025-12-03 22:16:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:02.499218386 +0000 UTC m=+2119.241928981" watchObservedRunningTime="2025-12-03 22:17:02.587911428 +0000 UTC m=+2119.330622023" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.590570 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.593569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvvwt\" (UniqueName: \"kubernetes.io/projected/4382c9dd-8a2d-4634-b9de-0796a994d491-kube-api-access-qvvwt\") pod \"barbican-worker-78449cf9c5-v4vnx\" (UID: \"4382c9dd-8a2d-4634-b9de-0796a994d491\") " pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.601753 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thtjg\" (UniqueName: \"kubernetes.io/projected/0e740246-a3e9-4692-8888-e12768f0573e-kube-api-access-thtjg\") pod \"barbican-keystone-listener-9cc86b6b-6xlfj\" (UID: \"0e740246-a3e9-4692-8888-e12768f0573e\") " pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.605530 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-srvlb"] Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.613235 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-666d579d9-hnzld" podStartSLOduration=8.613222985 podStartE2EDuration="8.613222985s" podCreationTimestamp="2025-12-03 22:16:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:02.55093785 +0000 UTC m=+2119.293648445" watchObservedRunningTime="2025-12-03 22:17:02.613222985 +0000 UTC m=+2119.355933580" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.661859 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.661921 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.661999 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfnbq\" (UniqueName: \"kubernetes.io/projected/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-kube-api-access-tfnbq\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662039 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-combined-ca-bundle\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662073 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-config\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662105 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-logs\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662140 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data-custom\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662155 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662173 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j8ph\" (UniqueName: \"kubernetes.io/projected/1812c74c-bfdd-4707-a57e-c2ae44813d19-kube-api-access-8j8ph\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662196 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662228 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.662429 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-78449cf9c5-v4vnx" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.679940 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765443 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765555 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfnbq\" (UniqueName: \"kubernetes.io/projected/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-kube-api-access-tfnbq\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765602 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-combined-ca-bundle\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765627 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-config\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765668 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-logs\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765698 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data-custom\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765716 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765742 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j8ph\" (UniqueName: \"kubernetes.io/projected/1812c74c-bfdd-4707-a57e-c2ae44813d19-kube-api-access-8j8ph\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765775 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.765801 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.766110 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.768147 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.768780 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-logs\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.769545 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-config\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.769580 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.774995 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-combined-ca-bundle\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.775783 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.783093 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data-custom\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.783770 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfnbq\" (UniqueName: \"kubernetes.io/projected/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-kube-api-access-tfnbq\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.786930 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.789169 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j8ph\" (UniqueName: \"kubernetes.io/projected/1812c74c-bfdd-4707-a57e-c2ae44813d19-kube-api-access-8j8ph\") pod \"dnsmasq-dns-848cf88cfc-srvlb\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.792310 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data\") pod \"barbican-api-7f9c554bf8-zrldg\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.797575 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:02 crc kubenswrapper[4715]: I1203 22:17:02.989292 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:03 crc kubenswrapper[4715]: I1203 22:17:03.005948 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:03 crc kubenswrapper[4715]: I1203 22:17:03.257993 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-78449cf9c5-v4vnx"] Dec 03 22:17:03 crc kubenswrapper[4715]: I1203 22:17:03.401199 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-9cc86b6b-6xlfj"] Dec 03 22:17:03 crc kubenswrapper[4715]: I1203 22:17:03.410200 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78449cf9c5-v4vnx" event={"ID":"4382c9dd-8a2d-4634-b9de-0796a994d491","Type":"ContainerStarted","Data":"74dbb27f4b2a7524fd7b8db879b362a678105e5be944dd6bd9e9f3385c73e7e3"} Dec 03 22:17:03 crc kubenswrapper[4715]: I1203 22:17:03.532024 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-76bcbf7d4d-l7qmc"] Dec 03 22:17:03 crc kubenswrapper[4715]: W1203 22:17:03.555176 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod954b69af_5e1b_4eec_a783_1cbe28e7c0ca.slice/crio-26bf8b3b2fa72b5f436f3df41b5d1315746f371bd916364542f4e8c04f4c7df5 WatchSource:0}: Error finding container 26bf8b3b2fa72b5f436f3df41b5d1315746f371bd916364542f4e8c04f4c7df5: Status 404 returned error can't find the container with id 26bf8b3b2fa72b5f436f3df41b5d1315746f371bd916364542f4e8c04f4c7df5 Dec 03 22:17:03 crc kubenswrapper[4715]: W1203 22:17:03.667851 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7e6cd7d_b1f9_4be3_b611_4efb245838bc.slice/crio-5b8b96ca47fa76f111a51a630d594ee307c08316fc66bc49dc32c4e330cb0046 WatchSource:0}: Error finding container 5b8b96ca47fa76f111a51a630d594ee307c08316fc66bc49dc32c4e330cb0046: Status 404 returned error can't find the container with id 5b8b96ca47fa76f111a51a630d594ee307c08316fc66bc49dc32c4e330cb0046 Dec 03 22:17:03 crc kubenswrapper[4715]: I1203 22:17:03.669248 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f9c554bf8-zrldg"] Dec 03 22:17:03 crc kubenswrapper[4715]: I1203 22:17:03.755775 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-srvlb"] Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.422269 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" event={"ID":"0e740246-a3e9-4692-8888-e12768f0573e","Type":"ContainerStarted","Data":"391db801e8a44cdab83ce0bb90397e346d1ad5b86181427c175abc0b65bf759b"} Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.427569 4715 generic.go:334] "Generic (PLEG): container finished" podID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerID="f7b2eead7bfb1f642128f6458ee351ce9acd51f97d015fa7e884738fd8f031e1" exitCode=0 Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.427621 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" event={"ID":"1812c74c-bfdd-4707-a57e-c2ae44813d19","Type":"ContainerDied","Data":"f7b2eead7bfb1f642128f6458ee351ce9acd51f97d015fa7e884738fd8f031e1"} Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.427644 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" event={"ID":"1812c74c-bfdd-4707-a57e-c2ae44813d19","Type":"ContainerStarted","Data":"416c3ed228f1152a5b24a159dc38130a969efb9a1d7af78258f25fb123b64f73"} Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.438159 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f9c554bf8-zrldg" event={"ID":"a7e6cd7d-b1f9-4be3-b611-4efb245838bc","Type":"ContainerStarted","Data":"5e2f2776663015bdd94f6fe6f38549d2c131e9bf5ac3ac0193d027e67d5919d5"} Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.438563 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f9c554bf8-zrldg" event={"ID":"a7e6cd7d-b1f9-4be3-b611-4efb245838bc","Type":"ContainerStarted","Data":"5b8b96ca47fa76f111a51a630d594ee307c08316fc66bc49dc32c4e330cb0046"} Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.448420 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" podUID="b197785c-154d-4e76-a57b-8310513175ba" containerName="dnsmasq-dns" containerID="cri-o://6a895988eed56b92a122486af19bc098702fedc2bc115239bf4827819508a394" gracePeriod=10 Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.449229 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-76bcbf7d4d-l7qmc" event={"ID":"954b69af-5e1b-4eec-a783-1cbe28e7c0ca","Type":"ContainerStarted","Data":"010e976b24fa3dd42b49bc8f4722780e3e61e54672ceb99a79ed17e5db7cc125"} Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.449255 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-76bcbf7d4d-l7qmc" event={"ID":"954b69af-5e1b-4eec-a783-1cbe28e7c0ca","Type":"ContainerStarted","Data":"26bf8b3b2fa72b5f436f3df41b5d1315746f371bd916364542f4e8c04f4c7df5"} Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.449267 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:04 crc kubenswrapper[4715]: I1203 22:17:04.493275 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-76bcbf7d4d-l7qmc" podStartSLOduration=2.493253279 podStartE2EDuration="2.493253279s" podCreationTimestamp="2025-12-03 22:17:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:04.466215556 +0000 UTC m=+2121.208926151" watchObservedRunningTime="2025-12-03 22:17:04.493253279 +0000 UTC m=+2121.235963884" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.459746 4715 generic.go:334] "Generic (PLEG): container finished" podID="b197785c-154d-4e76-a57b-8310513175ba" containerID="6a895988eed56b92a122486af19bc098702fedc2bc115239bf4827819508a394" exitCode=0 Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.459813 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" event={"ID":"b197785c-154d-4e76-a57b-8310513175ba","Type":"ContainerDied","Data":"6a895988eed56b92a122486af19bc098702fedc2bc115239bf4827819508a394"} Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.594167 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.684401 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-swift-storage-0\") pod \"b197785c-154d-4e76-a57b-8310513175ba\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.684590 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-644dm\" (UniqueName: \"kubernetes.io/projected/b197785c-154d-4e76-a57b-8310513175ba-kube-api-access-644dm\") pod \"b197785c-154d-4e76-a57b-8310513175ba\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.684745 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-nb\") pod \"b197785c-154d-4e76-a57b-8310513175ba\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.684832 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-sb\") pod \"b197785c-154d-4e76-a57b-8310513175ba\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.684978 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-config\") pod \"b197785c-154d-4e76-a57b-8310513175ba\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.685115 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-svc\") pod \"b197785c-154d-4e76-a57b-8310513175ba\" (UID: \"b197785c-154d-4e76-a57b-8310513175ba\") " Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.709102 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b197785c-154d-4e76-a57b-8310513175ba-kube-api-access-644dm" (OuterVolumeSpecName: "kube-api-access-644dm") pod "b197785c-154d-4e76-a57b-8310513175ba" (UID: "b197785c-154d-4e76-a57b-8310513175ba"). InnerVolumeSpecName "kube-api-access-644dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.739849 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b197785c-154d-4e76-a57b-8310513175ba" (UID: "b197785c-154d-4e76-a57b-8310513175ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.740963 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-config" (OuterVolumeSpecName: "config") pod "b197785c-154d-4e76-a57b-8310513175ba" (UID: "b197785c-154d-4e76-a57b-8310513175ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.741285 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b197785c-154d-4e76-a57b-8310513175ba" (UID: "b197785c-154d-4e76-a57b-8310513175ba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.754005 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b197785c-154d-4e76-a57b-8310513175ba" (UID: "b197785c-154d-4e76-a57b-8310513175ba"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.761269 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b197785c-154d-4e76-a57b-8310513175ba" (UID: "b197785c-154d-4e76-a57b-8310513175ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.787140 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.787170 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.787182 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-644dm\" (UniqueName: \"kubernetes.io/projected/b197785c-154d-4e76-a57b-8310513175ba-kube-api-access-644dm\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.787199 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.787209 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:05 crc kubenswrapper[4715]: I1203 22:17:05.787217 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b197785c-154d-4e76-a57b-8310513175ba-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.083336 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-dd5bb56d6-jb5n2"] Dec 03 22:17:06 crc kubenswrapper[4715]: E1203 22:17:06.083760 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b197785c-154d-4e76-a57b-8310513175ba" containerName="dnsmasq-dns" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.083778 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b197785c-154d-4e76-a57b-8310513175ba" containerName="dnsmasq-dns" Dec 03 22:17:06 crc kubenswrapper[4715]: E1203 22:17:06.083795 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b197785c-154d-4e76-a57b-8310513175ba" containerName="init" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.083801 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b197785c-154d-4e76-a57b-8310513175ba" containerName="init" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.083968 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b197785c-154d-4e76-a57b-8310513175ba" containerName="dnsmasq-dns" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.084880 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.086863 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.086900 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.099055 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-dd5bb56d6-jb5n2"] Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.195359 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-config-data\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.195407 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-config-data-custom\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.195454 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-combined-ca-bundle\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.195609 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-public-tls-certs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.195856 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9170a301-98ac-4953-ae87-4910bf0a466a-logs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.195896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lwjl\" (UniqueName: \"kubernetes.io/projected/9170a301-98ac-4953-ae87-4910bf0a466a-kube-api-access-9lwjl\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.195928 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-internal-tls-certs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.297703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-combined-ca-bundle\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.297752 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-public-tls-certs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.297819 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9170a301-98ac-4953-ae87-4910bf0a466a-logs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.297885 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lwjl\" (UniqueName: \"kubernetes.io/projected/9170a301-98ac-4953-ae87-4910bf0a466a-kube-api-access-9lwjl\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.297906 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-internal-tls-certs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.297997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-config-data\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.298031 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-config-data-custom\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.299102 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9170a301-98ac-4953-ae87-4910bf0a466a-logs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.303109 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-internal-tls-certs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.304170 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-combined-ca-bundle\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.304722 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-config-data\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.306119 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-public-tls-certs\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.307905 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9170a301-98ac-4953-ae87-4910bf0a466a-config-data-custom\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.332495 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lwjl\" (UniqueName: \"kubernetes.io/projected/9170a301-98ac-4953-ae87-4910bf0a466a-kube-api-access-9lwjl\") pod \"barbican-api-dd5bb56d6-jb5n2\" (UID: \"9170a301-98ac-4953-ae87-4910bf0a466a\") " pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.400419 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.470773 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" event={"ID":"1812c74c-bfdd-4707-a57e-c2ae44813d19","Type":"ContainerStarted","Data":"9ac3750e10043c097ca07f6f5b8dc3fce76ad776d5df17e0caf202c6ab68e576"} Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.471431 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.493553 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f9c554bf8-zrldg" event={"ID":"a7e6cd7d-b1f9-4be3-b611-4efb245838bc","Type":"ContainerStarted","Data":"cba38c86a3d4c38fafeabc21ade5e157bd44f053c133087223f32157a4646236"} Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.496568 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.496613 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.499423 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" event={"ID":"b197785c-154d-4e76-a57b-8310513175ba","Type":"ContainerDied","Data":"4c436b16395a56f7796a58be73abb04fd8548aa0d24d53acbca95f8f0ca635d9"} Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.499455 4715 scope.go:117] "RemoveContainer" containerID="6a895988eed56b92a122486af19bc098702fedc2bc115239bf4827819508a394" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.499593 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-7fj29" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.522275 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" podStartSLOduration=4.522256787 podStartE2EDuration="4.522256787s" podCreationTimestamp="2025-12-03 22:17:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:06.493325633 +0000 UTC m=+2123.236036228" watchObservedRunningTime="2025-12-03 22:17:06.522256787 +0000 UTC m=+2123.264967382" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.553159 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7f9c554bf8-zrldg" podStartSLOduration=4.553138112 podStartE2EDuration="4.553138112s" podCreationTimestamp="2025-12-03 22:17:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:06.539921739 +0000 UTC m=+2123.282632364" watchObservedRunningTime="2025-12-03 22:17:06.553138112 +0000 UTC m=+2123.295848707" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.577306 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-7fj29"] Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.586866 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-7fj29"] Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.588579 4715 scope.go:117] "RemoveContainer" containerID="9179ccf2edf7b5df92e12172e0eaddac8184a42ab9f884bd8d60549d991811c9" Dec 03 22:17:06 crc kubenswrapper[4715]: I1203 22:17:06.871929 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-dd5bb56d6-jb5n2"] Dec 03 22:17:07 crc kubenswrapper[4715]: I1203 22:17:07.531807 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd5bb56d6-jb5n2" event={"ID":"9170a301-98ac-4953-ae87-4910bf0a466a","Type":"ContainerStarted","Data":"a2fe7796f54de4ca15f0b657eadd2e5ad379024dd4f9d7185d63437768f3fdc2"} Dec 03 22:17:07 crc kubenswrapper[4715]: I1203 22:17:07.659708 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b197785c-154d-4e76-a57b-8310513175ba" path="/var/lib/kubelet/pods/b197785c-154d-4e76-a57b-8310513175ba/volumes" Dec 03 22:17:08 crc kubenswrapper[4715]: I1203 22:17:08.316956 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-74cd84ccf4-k9bcn" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 22:17:08 crc kubenswrapper[4715]: I1203 22:17:08.513237 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-656478594d-5tbrm" podUID="6c7cfacf-a283-45d3-982d-b28e62a9146b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Dec 03 22:17:13 crc kubenswrapper[4715]: I1203 22:17:13.007730 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:13 crc kubenswrapper[4715]: I1203 22:17:13.090664 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zf69s"] Dec 03 22:17:13 crc kubenswrapper[4715]: I1203 22:17:13.090887 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" podUID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerName="dnsmasq-dns" containerID="cri-o://aed158e03dfae84d343df09491228128b814863fd73cf6b514deb3ccef7a9767" gracePeriod=10 Dec 03 22:17:13 crc kubenswrapper[4715]: I1203 22:17:13.583983 4715 generic.go:334] "Generic (PLEG): container finished" podID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerID="aed158e03dfae84d343df09491228128b814863fd73cf6b514deb3ccef7a9767" exitCode=0 Dec 03 22:17:13 crc kubenswrapper[4715]: I1203 22:17:13.584046 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" event={"ID":"e05e03dd-8c4f-423c-bfba-db6d10fef841","Type":"ContainerDied","Data":"aed158e03dfae84d343df09491228128b814863fd73cf6b514deb3ccef7a9767"} Dec 03 22:17:14 crc kubenswrapper[4715]: I1203 22:17:14.439646 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" podUID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Dec 03 22:17:14 crc kubenswrapper[4715]: I1203 22:17:14.542071 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7f9c554bf8-zrldg" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 22:17:14 crc kubenswrapper[4715]: I1203 22:17:14.550280 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7f9c554bf8-zrldg" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 22:17:14 crc kubenswrapper[4715]: I1203 22:17:14.551519 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7f9c554bf8-zrldg" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 22:17:17 crc kubenswrapper[4715]: E1203 22:17:17.627372 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Dec 03 22:17:17 crc kubenswrapper[4715]: E1203 22:17:17.628038 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2lp2b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(72b03167-1df1-47c3-94c7-e0f13b65f972): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 22:17:17 crc kubenswrapper[4715]: E1203 22:17:17.629349 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" Dec 03 22:17:17 crc kubenswrapper[4715]: I1203 22:17:17.659265 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" event={"ID":"e05e03dd-8c4f-423c-bfba-db6d10fef841","Type":"ContainerDied","Data":"d8b89711705db7b5532bbd0e0103a21a9551fe9b9f807bee73c1c554dc81383a"} Dec 03 22:17:17 crc kubenswrapper[4715]: I1203 22:17:17.659340 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8b89711705db7b5532bbd0e0103a21a9551fe9b9f807bee73c1c554dc81383a" Dec 03 22:17:17 crc kubenswrapper[4715]: I1203 22:17:17.913142 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.055738 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.057010 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-config\") pod \"e05e03dd-8c4f-423c-bfba-db6d10fef841\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.057122 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grj5v\" (UniqueName: \"kubernetes.io/projected/e05e03dd-8c4f-423c-bfba-db6d10fef841-kube-api-access-grj5v\") pod \"e05e03dd-8c4f-423c-bfba-db6d10fef841\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.057273 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-sb\") pod \"e05e03dd-8c4f-423c-bfba-db6d10fef841\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.057367 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-svc\") pod \"e05e03dd-8c4f-423c-bfba-db6d10fef841\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.057399 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-nb\") pod \"e05e03dd-8c4f-423c-bfba-db6d10fef841\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.057661 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-swift-storage-0\") pod \"e05e03dd-8c4f-423c-bfba-db6d10fef841\" (UID: \"e05e03dd-8c4f-423c-bfba-db6d10fef841\") " Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.076889 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e05e03dd-8c4f-423c-bfba-db6d10fef841-kube-api-access-grj5v" (OuterVolumeSpecName: "kube-api-access-grj5v") pod "e05e03dd-8c4f-423c-bfba-db6d10fef841" (UID: "e05e03dd-8c4f-423c-bfba-db6d10fef841"). InnerVolumeSpecName "kube-api-access-grj5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.152276 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e05e03dd-8c4f-423c-bfba-db6d10fef841" (UID: "e05e03dd-8c4f-423c-bfba-db6d10fef841"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.162772 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.162800 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grj5v\" (UniqueName: \"kubernetes.io/projected/e05e03dd-8c4f-423c-bfba-db6d10fef841-kube-api-access-grj5v\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.197946 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e05e03dd-8c4f-423c-bfba-db6d10fef841" (UID: "e05e03dd-8c4f-423c-bfba-db6d10fef841"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.206171 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e05e03dd-8c4f-423c-bfba-db6d10fef841" (UID: "e05e03dd-8c4f-423c-bfba-db6d10fef841"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.206545 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e05e03dd-8c4f-423c-bfba-db6d10fef841" (UID: "e05e03dd-8c4f-423c-bfba-db6d10fef841"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.220085 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-config" (OuterVolumeSpecName: "config") pod "e05e03dd-8c4f-423c-bfba-db6d10fef841" (UID: "e05e03dd-8c4f-423c-bfba-db6d10fef841"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.266347 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.266381 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.266395 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.266406 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e05e03dd-8c4f-423c-bfba-db6d10fef841-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.650844 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd5bb56d6-jb5n2" event={"ID":"9170a301-98ac-4953-ae87-4910bf0a466a","Type":"ContainerStarted","Data":"506a422b111159ed69e6e573aaeffaef79f924ce56dba194bd61ea79dc021656"} Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.650885 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-dd5bb56d6-jb5n2" event={"ID":"9170a301-98ac-4953-ae87-4910bf0a466a","Type":"ContainerStarted","Data":"ab7958b6387fbc9db43725045164fe6a537ff9941f55d1734557dcc14f4f427e"} Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.651049 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.651083 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.654770 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78449cf9c5-v4vnx" event={"ID":"4382c9dd-8a2d-4634-b9de-0796a994d491","Type":"ContainerStarted","Data":"be851469f3029c0bc082acc4345f87f68dfe862341b8962aa6fe11d112927e49"} Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.654822 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-78449cf9c5-v4vnx" event={"ID":"4382c9dd-8a2d-4634-b9de-0796a994d491","Type":"ContainerStarted","Data":"bcbba46265b4b7b33bee11f4e669b7a76afd22e45007fe3b0abcae01198c2ec4"} Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.660668 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-zf69s" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.660709 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" event={"ID":"0e740246-a3e9-4692-8888-e12768f0573e","Type":"ContainerStarted","Data":"37e01c3fb6e5028b29550772028c497dc812f07aa9ea52b5afb58f0c671874da"} Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.660742 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" event={"ID":"0e740246-a3e9-4692-8888-e12768f0573e","Type":"ContainerStarted","Data":"c4da108ad65145333e2f38814e9b76fce41a7434053002557e6bcb13b4fa8dea"} Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.660847 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerName="ceilometer-notification-agent" containerID="cri-o://0239ba4c5711c4500c470cdc0c11f24c683374f0d887c7aa118733dae3186d74" gracePeriod=30 Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.660956 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerName="sg-core" containerID="cri-o://126f2d956398730e33ce6940fe54f6d10ae8f579570da1802dd5231d55ff29a9" gracePeriod=30 Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.674664 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-dd5bb56d6-jb5n2" podStartSLOduration=12.674648866 podStartE2EDuration="12.674648866s" podCreationTimestamp="2025-12-03 22:17:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:18.667012381 +0000 UTC m=+2135.409722976" watchObservedRunningTime="2025-12-03 22:17:18.674648866 +0000 UTC m=+2135.417359451" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.706861 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-9cc86b6b-6xlfj" podStartSLOduration=2.518572036 podStartE2EDuration="16.706846167s" podCreationTimestamp="2025-12-03 22:17:02 +0000 UTC" firstStartedPulling="2025-12-03 22:17:03.409642222 +0000 UTC m=+2120.152352817" lastFinishedPulling="2025-12-03 22:17:17.597916353 +0000 UTC m=+2134.340626948" observedRunningTime="2025-12-03 22:17:18.70625431 +0000 UTC m=+2135.448964905" watchObservedRunningTime="2025-12-03 22:17:18.706846167 +0000 UTC m=+2135.449556762" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.730105 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-78449cf9c5-v4vnx" podStartSLOduration=2.378714666 podStartE2EDuration="16.730087428s" podCreationTimestamp="2025-12-03 22:17:02 +0000 UTC" firstStartedPulling="2025-12-03 22:17:03.293625319 +0000 UTC m=+2120.036335904" lastFinishedPulling="2025-12-03 22:17:17.644998031 +0000 UTC m=+2134.387708666" observedRunningTime="2025-12-03 22:17:18.726566684 +0000 UTC m=+2135.469277279" watchObservedRunningTime="2025-12-03 22:17:18.730087428 +0000 UTC m=+2135.472798023" Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.750621 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zf69s"] Dec 03 22:17:18 crc kubenswrapper[4715]: I1203 22:17:18.759738 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-zf69s"] Dec 03 22:17:19 crc kubenswrapper[4715]: I1203 22:17:19.647253 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e05e03dd-8c4f-423c-bfba-db6d10fef841" path="/var/lib/kubelet/pods/e05e03dd-8c4f-423c-bfba-db6d10fef841/volumes" Dec 03 22:17:19 crc kubenswrapper[4715]: I1203 22:17:19.672791 4715 generic.go:334] "Generic (PLEG): container finished" podID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerID="126f2d956398730e33ce6940fe54f6d10ae8f579570da1802dd5231d55ff29a9" exitCode=2 Dec 03 22:17:19 crc kubenswrapper[4715]: I1203 22:17:19.672854 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72b03167-1df1-47c3-94c7-e0f13b65f972","Type":"ContainerDied","Data":"126f2d956398730e33ce6940fe54f6d10ae8f579570da1802dd5231d55ff29a9"} Dec 03 22:17:19 crc kubenswrapper[4715]: I1203 22:17:19.678634 4715 generic.go:334] "Generic (PLEG): container finished" podID="5ec32c20-9117-4eca-9b73-a7f8b84677da" containerID="16fd66c726542fba27a9cf3c4da98d61f54d7099995437d0013b10cd48184d2a" exitCode=0 Dec 03 22:17:19 crc kubenswrapper[4715]: I1203 22:17:19.679819 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p4mjk" event={"ID":"5ec32c20-9117-4eca-9b73-a7f8b84677da","Type":"ContainerDied","Data":"16fd66c726542fba27a9cf3c4da98d61f54d7099995437d0013b10cd48184d2a"} Dec 03 22:17:19 crc kubenswrapper[4715]: I1203 22:17:19.776787 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:20 crc kubenswrapper[4715]: I1203 22:17:20.574270 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:17:20 crc kubenswrapper[4715]: I1203 22:17:20.691004 4715 generic.go:334] "Generic (PLEG): container finished" podID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerID="0239ba4c5711c4500c470cdc0c11f24c683374f0d887c7aa118733dae3186d74" exitCode=0 Dec 03 22:17:20 crc kubenswrapper[4715]: I1203 22:17:20.691063 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72b03167-1df1-47c3-94c7-e0f13b65f972","Type":"ContainerDied","Data":"0239ba4c5711c4500c470cdc0c11f24c683374f0d887c7aa118733dae3186d74"} Dec 03 22:17:20 crc kubenswrapper[4715]: I1203 22:17:20.750371 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:17:20 crc kubenswrapper[4715]: I1203 22:17:20.980606 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.088397 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.130084 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-scripts\") pod \"72b03167-1df1-47c3-94c7-e0f13b65f972\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.130141 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-sg-core-conf-yaml\") pod \"72b03167-1df1-47c3-94c7-e0f13b65f972\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.130168 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-config-data\") pod \"72b03167-1df1-47c3-94c7-e0f13b65f972\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.130274 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-log-httpd\") pod \"72b03167-1df1-47c3-94c7-e0f13b65f972\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.130302 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-run-httpd\") pod \"72b03167-1df1-47c3-94c7-e0f13b65f972\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.130331 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2lp2b\" (UniqueName: \"kubernetes.io/projected/72b03167-1df1-47c3-94c7-e0f13b65f972-kube-api-access-2lp2b\") pod \"72b03167-1df1-47c3-94c7-e0f13b65f972\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.130410 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-combined-ca-bundle\") pod \"72b03167-1df1-47c3-94c7-e0f13b65f972\" (UID: \"72b03167-1df1-47c3-94c7-e0f13b65f972\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.131543 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "72b03167-1df1-47c3-94c7-e0f13b65f972" (UID: "72b03167-1df1-47c3-94c7-e0f13b65f972"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.131662 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "72b03167-1df1-47c3-94c7-e0f13b65f972" (UID: "72b03167-1df1-47c3-94c7-e0f13b65f972"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.136520 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-scripts" (OuterVolumeSpecName: "scripts") pod "72b03167-1df1-47c3-94c7-e0f13b65f972" (UID: "72b03167-1df1-47c3-94c7-e0f13b65f972"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.136703 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b03167-1df1-47c3-94c7-e0f13b65f972-kube-api-access-2lp2b" (OuterVolumeSpecName: "kube-api-access-2lp2b") pod "72b03167-1df1-47c3-94c7-e0f13b65f972" (UID: "72b03167-1df1-47c3-94c7-e0f13b65f972"). InnerVolumeSpecName "kube-api-access-2lp2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.158180 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72b03167-1df1-47c3-94c7-e0f13b65f972" (UID: "72b03167-1df1-47c3-94c7-e0f13b65f972"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.163304 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-config-data" (OuterVolumeSpecName: "config-data") pod "72b03167-1df1-47c3-94c7-e0f13b65f972" (UID: "72b03167-1df1-47c3-94c7-e0f13b65f972"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.163463 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "72b03167-1df1-47c3-94c7-e0f13b65f972" (UID: "72b03167-1df1-47c3-94c7-e0f13b65f972"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.231856 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-combined-ca-bundle\") pod \"5ec32c20-9117-4eca-9b73-a7f8b84677da\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.231904 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec32c20-9117-4eca-9b73-a7f8b84677da-etc-machine-id\") pod \"5ec32c20-9117-4eca-9b73-a7f8b84677da\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.231959 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-db-sync-config-data\") pod \"5ec32c20-9117-4eca-9b73-a7f8b84677da\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232010 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9t27x\" (UniqueName: \"kubernetes.io/projected/5ec32c20-9117-4eca-9b73-a7f8b84677da-kube-api-access-9t27x\") pod \"5ec32c20-9117-4eca-9b73-a7f8b84677da\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232054 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-config-data\") pod \"5ec32c20-9117-4eca-9b73-a7f8b84677da\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232129 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-scripts\") pod \"5ec32c20-9117-4eca-9b73-a7f8b84677da\" (UID: \"5ec32c20-9117-4eca-9b73-a7f8b84677da\") " Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232683 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232697 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/72b03167-1df1-47c3-94c7-e0f13b65f972-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232708 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2lp2b\" (UniqueName: \"kubernetes.io/projected/72b03167-1df1-47c3-94c7-e0f13b65f972-kube-api-access-2lp2b\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232721 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232732 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232743 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.232756 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72b03167-1df1-47c3-94c7-e0f13b65f972-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.233688 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5ec32c20-9117-4eca-9b73-a7f8b84677da-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5ec32c20-9117-4eca-9b73-a7f8b84677da" (UID: "5ec32c20-9117-4eca-9b73-a7f8b84677da"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.237413 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-scripts" (OuterVolumeSpecName: "scripts") pod "5ec32c20-9117-4eca-9b73-a7f8b84677da" (UID: "5ec32c20-9117-4eca-9b73-a7f8b84677da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.253056 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec32c20-9117-4eca-9b73-a7f8b84677da-kube-api-access-9t27x" (OuterVolumeSpecName: "kube-api-access-9t27x") pod "5ec32c20-9117-4eca-9b73-a7f8b84677da" (UID: "5ec32c20-9117-4eca-9b73-a7f8b84677da"). InnerVolumeSpecName "kube-api-access-9t27x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.253443 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5ec32c20-9117-4eca-9b73-a7f8b84677da" (UID: "5ec32c20-9117-4eca-9b73-a7f8b84677da"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.259196 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ec32c20-9117-4eca-9b73-a7f8b84677da" (UID: "5ec32c20-9117-4eca-9b73-a7f8b84677da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.282849 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-config-data" (OuterVolumeSpecName: "config-data") pod "5ec32c20-9117-4eca-9b73-a7f8b84677da" (UID: "5ec32c20-9117-4eca-9b73-a7f8b84677da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.334482 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.334531 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5ec32c20-9117-4eca-9b73-a7f8b84677da-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.334540 4715 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.334549 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9t27x\" (UniqueName: \"kubernetes.io/projected/5ec32c20-9117-4eca-9b73-a7f8b84677da-kube-api-access-9t27x\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.334562 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.334570 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ec32c20-9117-4eca-9b73-a7f8b84677da-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.714126 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"72b03167-1df1-47c3-94c7-e0f13b65f972","Type":"ContainerDied","Data":"8e1527e2a13e8a9a7188b887d9741cfc81f3bb52816d223d553d52ba2fa0d068"} Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.714159 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.714191 4715 scope.go:117] "RemoveContainer" containerID="126f2d956398730e33ce6940fe54f6d10ae8f579570da1802dd5231d55ff29a9" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.725010 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-p4mjk" event={"ID":"5ec32c20-9117-4eca-9b73-a7f8b84677da","Type":"ContainerDied","Data":"b96444269274e802b8f5aba353bfcf7156bc2e86cdd2fcbd3130f4f86f71ac13"} Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.725060 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b96444269274e802b8f5aba353bfcf7156bc2e86cdd2fcbd3130f4f86f71ac13" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.725123 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-p4mjk" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.745597 4715 scope.go:117] "RemoveContainer" containerID="0239ba4c5711c4500c470cdc0c11f24c683374f0d887c7aa118733dae3186d74" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.786560 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.804621 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818000 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:21 crc kubenswrapper[4715]: E1203 22:17:21.818434 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec32c20-9117-4eca-9b73-a7f8b84677da" containerName="cinder-db-sync" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818450 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec32c20-9117-4eca-9b73-a7f8b84677da" containerName="cinder-db-sync" Dec 03 22:17:21 crc kubenswrapper[4715]: E1203 22:17:21.818466 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerName="sg-core" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818472 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerName="sg-core" Dec 03 22:17:21 crc kubenswrapper[4715]: E1203 22:17:21.818488 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerName="ceilometer-notification-agent" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818497 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerName="ceilometer-notification-agent" Dec 03 22:17:21 crc kubenswrapper[4715]: E1203 22:17:21.818523 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerName="init" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818531 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerName="init" Dec 03 22:17:21 crc kubenswrapper[4715]: E1203 22:17:21.818566 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerName="dnsmasq-dns" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818573 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerName="dnsmasq-dns" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818824 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerName="sg-core" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818844 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec32c20-9117-4eca-9b73-a7f8b84677da" containerName="cinder-db-sync" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818900 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e05e03dd-8c4f-423c-bfba-db6d10fef841" containerName="dnsmasq-dns" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.818912 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" containerName="ceilometer-notification-agent" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.822047 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.826948 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.827030 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.832699 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.948654 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-scripts\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.948699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h65nt\" (UniqueName: \"kubernetes.io/projected/3b11ee3d-7aea-416f-b035-1ad5efe949e2-kube-api-access-h65nt\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.948727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.948777 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-log-httpd\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.948794 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.948815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-run-httpd\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:21 crc kubenswrapper[4715]: I1203 22:17:21.948869 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-config-data\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.002140 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.005288 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.015067 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.015254 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-wb6bd" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.015586 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.015618 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.018971 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050073 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h65nt\" (UniqueName: \"kubernetes.io/projected/3b11ee3d-7aea-416f-b035-1ad5efe949e2-kube-api-access-h65nt\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050128 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050164 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn6nz\" (UniqueName: \"kubernetes.io/projected/905c7551-4b61-4cbe-a440-9a786e863ed1-kube-api-access-cn6nz\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050254 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-log-httpd\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050277 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050309 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-run-httpd\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050341 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050383 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-scripts\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050423 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050462 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-config-data\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050493 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050555 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/905c7551-4b61-4cbe-a440-9a786e863ed1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.050613 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-scripts\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.053103 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-run-httpd\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.054977 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-log-httpd\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.057091 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.065601 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-scripts\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.067304 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.069439 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-config-data\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.092199 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-fhfjs"] Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.129964 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.145561 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h65nt\" (UniqueName: \"kubernetes.io/projected/3b11ee3d-7aea-416f-b035-1ad5efe949e2-kube-api-access-h65nt\") pod \"ceilometer-0\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.152293 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.158362 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn6nz\" (UniqueName: \"kubernetes.io/projected/905c7551-4b61-4cbe-a440-9a786e863ed1-kube-api-access-cn6nz\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.158544 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.158630 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-scripts\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.158666 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.158713 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.158795 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/905c7551-4b61-4cbe-a440-9a786e863ed1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.159097 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/905c7551-4b61-4cbe-a440-9a786e863ed1-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.160550 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-fhfjs"] Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.164240 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.167248 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-scripts\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.208440 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.208604 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.212778 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn6nz\" (UniqueName: \"kubernetes.io/projected/905c7551-4b61-4cbe-a440-9a786e863ed1-kube-api-access-cn6nz\") pod \"cinder-scheduler-0\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.265396 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.266900 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.266984 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5jzp\" (UniqueName: \"kubernetes.io/projected/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-kube-api-access-p5jzp\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.267023 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.267062 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-config\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.267086 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.267170 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-svc\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.323986 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.325905 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.335091 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.363223 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.369877 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.369919 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48f63067-3d5d-4df6-bb37-cb03a29bdc96-etc-machine-id\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370009 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmkp8\" (UniqueName: \"kubernetes.io/projected/48f63067-3d5d-4df6-bb37-cb03a29bdc96-kube-api-access-cmkp8\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370035 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370078 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5jzp\" (UniqueName: \"kubernetes.io/projected/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-kube-api-access-p5jzp\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370115 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370141 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-scripts\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370189 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-config\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370213 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370312 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370367 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data-custom\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370390 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48f63067-3d5d-4df6-bb37-cb03a29bdc96-logs\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.370431 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-svc\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.379175 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.379575 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.380000 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-config\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.381248 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.397295 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-svc\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.406772 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5jzp\" (UniqueName: \"kubernetes.io/projected/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-kube-api-access-p5jzp\") pod \"dnsmasq-dns-6578955fd5-fhfjs\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.474879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-scripts\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.474965 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.474997 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data-custom\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.475014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48f63067-3d5d-4df6-bb37-cb03a29bdc96-logs\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.475041 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48f63067-3d5d-4df6-bb37-cb03a29bdc96-etc-machine-id\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.475087 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmkp8\" (UniqueName: \"kubernetes.io/projected/48f63067-3d5d-4df6-bb37-cb03a29bdc96-kube-api-access-cmkp8\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.475104 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.487413 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48f63067-3d5d-4df6-bb37-cb03a29bdc96-etc-machine-id\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.487996 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48f63067-3d5d-4df6-bb37-cb03a29bdc96-logs\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.489611 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.498327 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data-custom\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.505155 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.509944 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-scripts\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.544074 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmkp8\" (UniqueName: \"kubernetes.io/projected/48f63067-3d5d-4df6-bb37-cb03a29bdc96-kube-api-access-cmkp8\") pod \"cinder-api-0\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.586337 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.657053 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.660148 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-656478594d-5tbrm" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.728005 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74cd84ccf4-k9bcn"] Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.728277 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74cd84ccf4-k9bcn" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon-log" containerID="cri-o://706dc959898b0c6a666acb6b020a24e81832e5be7099c366f91458474d2b0618" gracePeriod=30 Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.728686 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-74cd84ccf4-k9bcn" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" containerID="cri-o://4979bdb653c8db320223f4fb81a9d23bee0c19d8f6be47165c7af4d53c639d1f" gracePeriod=30 Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.741269 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-74cd84ccf4-k9bcn" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": EOF" Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.873991 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:22 crc kubenswrapper[4715]: I1203 22:17:22.966665 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:17:23 crc kubenswrapper[4715]: I1203 22:17:23.017706 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:23 crc kubenswrapper[4715]: I1203 22:17:23.212178 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-fhfjs"] Dec 03 22:17:23 crc kubenswrapper[4715]: I1203 22:17:23.349078 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:23 crc kubenswrapper[4715]: W1203 22:17:23.369975 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48f63067_3d5d_4df6_bb37_cb03a29bdc96.slice/crio-2ef096f37cbcf9cbb409b6d40d06a77f80e89d38c5e7c88660a1c7f8028447df WatchSource:0}: Error finding container 2ef096f37cbcf9cbb409b6d40d06a77f80e89d38c5e7c88660a1c7f8028447df: Status 404 returned error can't find the container with id 2ef096f37cbcf9cbb409b6d40d06a77f80e89d38c5e7c88660a1c7f8028447df Dec 03 22:17:23 crc kubenswrapper[4715]: I1203 22:17:23.654685 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b03167-1df1-47c3-94c7-e0f13b65f972" path="/var/lib/kubelet/pods/72b03167-1df1-47c3-94c7-e0f13b65f972/volumes" Dec 03 22:17:23 crc kubenswrapper[4715]: I1203 22:17:23.805540 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" event={"ID":"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f","Type":"ContainerStarted","Data":"03407d6c823302ef91a508113cd5ce4e0020e62236ef70f659542a9701916a7c"} Dec 03 22:17:23 crc kubenswrapper[4715]: I1203 22:17:23.806851 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"905c7551-4b61-4cbe-a440-9a786e863ed1","Type":"ContainerStarted","Data":"10df1a48a9a81a3391c7d9f70df1a00a961aaa2a94a4cecdb7532784e58af744"} Dec 03 22:17:23 crc kubenswrapper[4715]: I1203 22:17:23.808213 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48f63067-3d5d-4df6-bb37-cb03a29bdc96","Type":"ContainerStarted","Data":"2ef096f37cbcf9cbb409b6d40d06a77f80e89d38c5e7c88660a1c7f8028447df"} Dec 03 22:17:23 crc kubenswrapper[4715]: I1203 22:17:23.812838 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerStarted","Data":"c4b44f0a3896c1c5685c7cc46c1dadbaa949c81720dbbc07c711fb9029c181e1"} Dec 03 22:17:24 crc kubenswrapper[4715]: I1203 22:17:24.383274 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:24 crc kubenswrapper[4715]: I1203 22:17:24.614442 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 22:17:24 crc kubenswrapper[4715]: I1203 22:17:24.615732 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 22:17:24 crc kubenswrapper[4715]: I1203 22:17:24.827713 4715 generic.go:334] "Generic (PLEG): container finished" podID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" containerID="bd19da69710577d4336bbfe6bd35bf47ed5e47784c2fffa80d530f10e427e6de" exitCode=0 Dec 03 22:17:24 crc kubenswrapper[4715]: I1203 22:17:24.827756 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" event={"ID":"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f","Type":"ContainerDied","Data":"bd19da69710577d4336bbfe6bd35bf47ed5e47784c2fffa80d530f10e427e6de"} Dec 03 22:17:24 crc kubenswrapper[4715]: I1203 22:17:24.844826 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48f63067-3d5d-4df6-bb37-cb03a29bdc96","Type":"ContainerStarted","Data":"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85"} Dec 03 22:17:24 crc kubenswrapper[4715]: I1203 22:17:24.982171 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8694d8458b-2x672" Dec 03 22:17:24 crc kubenswrapper[4715]: I1203 22:17:24.988800 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-8694d8458b-2x672" Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.141864 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-666d579d9-hnzld" Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.199157 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8c9bc67fb-47rvm"] Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.199365 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8c9bc67fb-47rvm" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerName="neutron-api" containerID="cri-o://f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e" gracePeriod=30 Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.199728 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-8c9bc67fb-47rvm" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerName="neutron-httpd" containerID="cri-o://ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b" gracePeriod=30 Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.887294 4715 generic.go:334] "Generic (PLEG): container finished" podID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerID="ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b" exitCode=0 Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.887390 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c9bc67fb-47rvm" event={"ID":"de1bec8a-6677-4be3-8722-5ec1bb336d6f","Type":"ContainerDied","Data":"ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b"} Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.890472 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerStarted","Data":"ca95a4cb86b25c3c32bc5f98fcc82f447b929f396069f8e09590cfe8a5a40ceb"} Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.892898 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" event={"ID":"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f","Type":"ContainerStarted","Data":"d1971a09d49c8130bc0a70f573319660f3e7e1990d43eafef19855a34f28f459"} Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.893029 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.902813 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerName="cinder-api-log" containerID="cri-o://8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85" gracePeriod=30 Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.903155 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerName="cinder-api" containerID="cri-o://161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54" gracePeriod=30 Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.903158 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48f63067-3d5d-4df6-bb37-cb03a29bdc96","Type":"ContainerStarted","Data":"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54"} Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.903233 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.931652 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" podStartSLOduration=3.9316252240000003 podStartE2EDuration="3.931625224s" podCreationTimestamp="2025-12-03 22:17:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:25.917906077 +0000 UTC m=+2142.660616672" watchObservedRunningTime="2025-12-03 22:17:25.931625224 +0000 UTC m=+2142.674335819" Dec 03 22:17:25 crc kubenswrapper[4715]: I1203 22:17:25.948131 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.948114934 podStartE2EDuration="3.948114934s" podCreationTimestamp="2025-12-03 22:17:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:25.936240997 +0000 UTC m=+2142.678951582" watchObservedRunningTime="2025-12-03 22:17:25.948114934 +0000 UTC m=+2142.690825529" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.580916 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.689582 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data\") pod \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.690197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmkp8\" (UniqueName: \"kubernetes.io/projected/48f63067-3d5d-4df6-bb37-cb03a29bdc96-kube-api-access-cmkp8\") pod \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.690306 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-combined-ca-bundle\") pod \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.690427 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48f63067-3d5d-4df6-bb37-cb03a29bdc96-etc-machine-id\") pod \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.690626 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-scripts\") pod \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.690726 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48f63067-3d5d-4df6-bb37-cb03a29bdc96-logs\") pod \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.690854 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data-custom\") pod \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\" (UID: \"48f63067-3d5d-4df6-bb37-cb03a29bdc96\") " Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.693959 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/48f63067-3d5d-4df6-bb37-cb03a29bdc96-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "48f63067-3d5d-4df6-bb37-cb03a29bdc96" (UID: "48f63067-3d5d-4df6-bb37-cb03a29bdc96"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.695116 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48f63067-3d5d-4df6-bb37-cb03a29bdc96-logs" (OuterVolumeSpecName: "logs") pod "48f63067-3d5d-4df6-bb37-cb03a29bdc96" (UID: "48f63067-3d5d-4df6-bb37-cb03a29bdc96"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.701633 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-scripts" (OuterVolumeSpecName: "scripts") pod "48f63067-3d5d-4df6-bb37-cb03a29bdc96" (UID: "48f63067-3d5d-4df6-bb37-cb03a29bdc96"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.701643 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48f63067-3d5d-4df6-bb37-cb03a29bdc96-kube-api-access-cmkp8" (OuterVolumeSpecName: "kube-api-access-cmkp8") pod "48f63067-3d5d-4df6-bb37-cb03a29bdc96" (UID: "48f63067-3d5d-4df6-bb37-cb03a29bdc96"). InnerVolumeSpecName "kube-api-access-cmkp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.716694 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "48f63067-3d5d-4df6-bb37-cb03a29bdc96" (UID: "48f63067-3d5d-4df6-bb37-cb03a29bdc96"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.741867 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data" (OuterVolumeSpecName: "config-data") pod "48f63067-3d5d-4df6-bb37-cb03a29bdc96" (UID: "48f63067-3d5d-4df6-bb37-cb03a29bdc96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.742745 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48f63067-3d5d-4df6-bb37-cb03a29bdc96" (UID: "48f63067-3d5d-4df6-bb37-cb03a29bdc96"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.792493 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.792770 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.792833 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmkp8\" (UniqueName: \"kubernetes.io/projected/48f63067-3d5d-4df6-bb37-cb03a29bdc96-kube-api-access-cmkp8\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.792886 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.792935 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/48f63067-3d5d-4df6-bb37-cb03a29bdc96-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.792981 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48f63067-3d5d-4df6-bb37-cb03a29bdc96-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.793034 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/48f63067-3d5d-4df6-bb37-cb03a29bdc96-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.911049 4715 generic.go:334] "Generic (PLEG): container finished" podID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerID="161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54" exitCode=0 Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.911084 4715 generic.go:334] "Generic (PLEG): container finished" podID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerID="8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85" exitCode=143 Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.911090 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.911136 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48f63067-3d5d-4df6-bb37-cb03a29bdc96","Type":"ContainerDied","Data":"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54"} Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.911165 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48f63067-3d5d-4df6-bb37-cb03a29bdc96","Type":"ContainerDied","Data":"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85"} Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.911174 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"48f63067-3d5d-4df6-bb37-cb03a29bdc96","Type":"ContainerDied","Data":"2ef096f37cbcf9cbb409b6d40d06a77f80e89d38c5e7c88660a1c7f8028447df"} Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.911189 4715 scope.go:117] "RemoveContainer" containerID="161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.915314 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerStarted","Data":"71132a3bc090ff4256672c9f66529c3304afd4b9efb8fc7179452076ef42610e"} Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.915426 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerStarted","Data":"128f27f9955d3da7ff215a33f665159a95a15070dbc5c6de235f1db3c70e9993"} Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.919861 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"905c7551-4b61-4cbe-a440-9a786e863ed1","Type":"ContainerStarted","Data":"e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61"} Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.920971 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"905c7551-4b61-4cbe-a440-9a786e863ed1","Type":"ContainerStarted","Data":"261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a"} Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.945667 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.664899638 podStartE2EDuration="5.945648911s" podCreationTimestamp="2025-12-03 22:17:21 +0000 UTC" firstStartedPulling="2025-12-03 22:17:23.009829897 +0000 UTC m=+2139.752540492" lastFinishedPulling="2025-12-03 22:17:25.29057917 +0000 UTC m=+2142.033289765" observedRunningTime="2025-12-03 22:17:26.941227233 +0000 UTC m=+2143.683937828" watchObservedRunningTime="2025-12-03 22:17:26.945648911 +0000 UTC m=+2143.688359506" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.974576 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.982749 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.995707 4715 scope.go:117] "RemoveContainer" containerID="8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.998357 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:26 crc kubenswrapper[4715]: E1203 22:17:26.998854 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerName="cinder-api-log" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.998870 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerName="cinder-api-log" Dec 03 22:17:26 crc kubenswrapper[4715]: E1203 22:17:26.998886 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerName="cinder-api" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.998893 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerName="cinder-api" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.999082 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerName="cinder-api-log" Dec 03 22:17:26 crc kubenswrapper[4715]: I1203 22:17:26.999106 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" containerName="cinder-api" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.000096 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.003175 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.003337 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.003370 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.011085 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.054622 4715 scope.go:117] "RemoveContainer" containerID="161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54" Dec 03 22:17:27 crc kubenswrapper[4715]: E1203 22:17:27.055397 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54\": container with ID starting with 161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54 not found: ID does not exist" containerID="161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.055439 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54"} err="failed to get container status \"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54\": rpc error: code = NotFound desc = could not find container \"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54\": container with ID starting with 161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54 not found: ID does not exist" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.055466 4715 scope.go:117] "RemoveContainer" containerID="8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85" Dec 03 22:17:27 crc kubenswrapper[4715]: E1203 22:17:27.058724 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85\": container with ID starting with 8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85 not found: ID does not exist" containerID="8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.058843 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85"} err="failed to get container status \"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85\": rpc error: code = NotFound desc = could not find container \"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85\": container with ID starting with 8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85 not found: ID does not exist" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.058931 4715 scope.go:117] "RemoveContainer" containerID="161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.062716 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54"} err="failed to get container status \"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54\": rpc error: code = NotFound desc = could not find container \"161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54\": container with ID starting with 161aa9d4bb0a6846c71bd2625ff6135de37d9eee639060b5a44ab89e23dc3c54 not found: ID does not exist" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.062851 4715 scope.go:117] "RemoveContainer" containerID="8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.066701 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85"} err="failed to get container status \"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85\": rpc error: code = NotFound desc = could not find container \"8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85\": container with ID starting with 8943727c2a11a2da579521d42b68dbd864cc914e3fd93240adac47fddadaae85 not found: ID does not exist" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.122221 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-74cd84ccf4-k9bcn" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:39408->10.217.0.144:8443: read: connection reset by peer" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.198497 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.198539 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.198568 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-scripts\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.198592 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-config-data\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.198614 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-logs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.198744 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqwkm\" (UniqueName: \"kubernetes.io/projected/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-kube-api-access-nqwkm\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.198868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-config-data-custom\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.198893 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.199020 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.266378 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.301049 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.301285 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.301362 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-scripts\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.301437 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-config-data\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.301598 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-logs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.301703 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqwkm\" (UniqueName: \"kubernetes.io/projected/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-kube-api-access-nqwkm\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.302021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-config-data-custom\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.302088 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.302183 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.302947 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-logs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.303616 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.315113 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.317062 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.319432 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-config-data\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.321132 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-config-data-custom\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.321379 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.322275 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-scripts\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.323310 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqwkm\" (UniqueName: \"kubernetes.io/projected/6cfc1c9d-e789-449f-a73d-a4aead3a03f1-kube-api-access-nqwkm\") pod \"cinder-api-0\" (UID: \"6cfc1c9d-e789-449f-a73d-a4aead3a03f1\") " pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.371525 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.653537 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48f63067-3d5d-4df6-bb37-cb03a29bdc96" path="/var/lib/kubelet/pods/48f63067-3d5d-4df6-bb37-cb03a29bdc96/volumes" Dec 03 22:17:27 crc kubenswrapper[4715]: W1203 22:17:27.871106 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cfc1c9d_e789_449f_a73d_a4aead3a03f1.slice/crio-b1291a8d5a53a3e788df15dd69c150d17638809daf29a9567fb0fd0ca0e09d03 WatchSource:0}: Error finding container b1291a8d5a53a3e788df15dd69c150d17638809daf29a9567fb0fd0ca0e09d03: Status 404 returned error can't find the container with id b1291a8d5a53a3e788df15dd69c150d17638809daf29a9567fb0fd0ca0e09d03 Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.880137 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.883197 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.947058 4715 generic.go:334] "Generic (PLEG): container finished" podID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerID="4979bdb653c8db320223f4fb81a9d23bee0c19d8f6be47165c7af4d53c639d1f" exitCode=0 Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.947185 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cd84ccf4-k9bcn" event={"ID":"ae09bc5f-d3e1-48c3-8145-611fd9827870","Type":"ContainerDied","Data":"4979bdb653c8db320223f4fb81a9d23bee0c19d8f6be47165c7af4d53c639d1f"} Dec 03 22:17:27 crc kubenswrapper[4715]: I1203 22:17:27.952073 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6cfc1c9d-e789-449f-a73d-a4aead3a03f1","Type":"ContainerStarted","Data":"b1291a8d5a53a3e788df15dd69c150d17638809daf29a9567fb0fd0ca0e09d03"} Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.256313 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-dd5bb56d6-jb5n2" Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.320756 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-74cd84ccf4-k9bcn" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.343145 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7f9c554bf8-zrldg"] Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.343465 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7f9c554bf8-zrldg" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api-log" containerID="cri-o://5e2f2776663015bdd94f6fe6f38549d2c131e9bf5ac3ac0193d027e67d5919d5" gracePeriod=30 Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.344059 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-7f9c554bf8-zrldg" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api" containerID="cri-o://cba38c86a3d4c38fafeabc21ade5e157bd44f053c133087223f32157a4646236" gracePeriod=30 Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.964242 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerStarted","Data":"fa615843941ce2dc9968739005e78812be531b42166aabb607eb92cb2f231dcc"} Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.964814 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.967969 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6cfc1c9d-e789-449f-a73d-a4aead3a03f1","Type":"ContainerStarted","Data":"b0addf269b1b2cbf99255e3d0556310384dac424f059d587f375f36521dbd3de"} Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.970614 4715 generic.go:334] "Generic (PLEG): container finished" podID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerID="5e2f2776663015bdd94f6fe6f38549d2c131e9bf5ac3ac0193d027e67d5919d5" exitCode=143 Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.971657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f9c554bf8-zrldg" event={"ID":"a7e6cd7d-b1f9-4be3-b611-4efb245838bc","Type":"ContainerDied","Data":"5e2f2776663015bdd94f6fe6f38549d2c131e9bf5ac3ac0193d027e67d5919d5"} Dec 03 22:17:28 crc kubenswrapper[4715]: I1203 22:17:28.990250 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.266942575 podStartE2EDuration="7.990230958s" podCreationTimestamp="2025-12-03 22:17:21 +0000 UTC" firstStartedPulling="2025-12-03 22:17:22.898222792 +0000 UTC m=+2139.640933377" lastFinishedPulling="2025-12-03 22:17:27.621511165 +0000 UTC m=+2144.364221760" observedRunningTime="2025-12-03 22:17:28.980624041 +0000 UTC m=+2145.723334646" watchObservedRunningTime="2025-12-03 22:17:28.990230958 +0000 UTC m=+2145.732941573" Dec 03 22:17:29 crc kubenswrapper[4715]: I1203 22:17:29.982270 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6cfc1c9d-e789-449f-a73d-a4aead3a03f1","Type":"ContainerStarted","Data":"57e995cccb10ee49fe37013722c513b798fdf616f6da461179e63204b119bd47"} Dec 03 22:17:29 crc kubenswrapper[4715]: I1203 22:17:29.982925 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.008953 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.008934331 podStartE2EDuration="4.008934331s" podCreationTimestamp="2025-12-03 22:17:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:29.999718345 +0000 UTC m=+2146.742428950" watchObservedRunningTime="2025-12-03 22:17:30.008934331 +0000 UTC m=+2146.751644936" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.691800 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.786497 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-ovndb-tls-certs\") pod \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.786926 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-httpd-config\") pod \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.787079 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-config\") pod \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.787157 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4ph5\" (UniqueName: \"kubernetes.io/projected/de1bec8a-6677-4be3-8722-5ec1bb336d6f-kube-api-access-z4ph5\") pod \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.787189 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-combined-ca-bundle\") pod \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\" (UID: \"de1bec8a-6677-4be3-8722-5ec1bb336d6f\") " Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.794927 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de1bec8a-6677-4be3-8722-5ec1bb336d6f-kube-api-access-z4ph5" (OuterVolumeSpecName: "kube-api-access-z4ph5") pod "de1bec8a-6677-4be3-8722-5ec1bb336d6f" (UID: "de1bec8a-6677-4be3-8722-5ec1bb336d6f"). InnerVolumeSpecName "kube-api-access-z4ph5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.795010 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "de1bec8a-6677-4be3-8722-5ec1bb336d6f" (UID: "de1bec8a-6677-4be3-8722-5ec1bb336d6f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.840627 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-config" (OuterVolumeSpecName: "config") pod "de1bec8a-6677-4be3-8722-5ec1bb336d6f" (UID: "de1bec8a-6677-4be3-8722-5ec1bb336d6f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.850693 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de1bec8a-6677-4be3-8722-5ec1bb336d6f" (UID: "de1bec8a-6677-4be3-8722-5ec1bb336d6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.878661 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "de1bec8a-6677-4be3-8722-5ec1bb336d6f" (UID: "de1bec8a-6677-4be3-8722-5ec1bb336d6f"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.889574 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.889614 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.889627 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.889643 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4ph5\" (UniqueName: \"kubernetes.io/projected/de1bec8a-6677-4be3-8722-5ec1bb336d6f-kube-api-access-z4ph5\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:30 crc kubenswrapper[4715]: I1203 22:17:30.889655 4715 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/de1bec8a-6677-4be3-8722-5ec1bb336d6f-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.003330 4715 generic.go:334] "Generic (PLEG): container finished" podID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerID="f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e" exitCode=0 Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.003437 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8c9bc67fb-47rvm" Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.003447 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c9bc67fb-47rvm" event={"ID":"de1bec8a-6677-4be3-8722-5ec1bb336d6f","Type":"ContainerDied","Data":"f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e"} Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.003578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8c9bc67fb-47rvm" event={"ID":"de1bec8a-6677-4be3-8722-5ec1bb336d6f","Type":"ContainerDied","Data":"b90af023126aeb6fe6fecba141c3ec5af87623d45986777aa4c6cd3a134ed90b"} Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.003619 4715 scope.go:117] "RemoveContainer" containerID="ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b" Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.038242 4715 scope.go:117] "RemoveContainer" containerID="f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e" Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.063764 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8c9bc67fb-47rvm"] Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.076527 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8c9bc67fb-47rvm"] Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.076755 4715 scope.go:117] "RemoveContainer" containerID="ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b" Dec 03 22:17:31 crc kubenswrapper[4715]: E1203 22:17:31.077220 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b\": container with ID starting with ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b not found: ID does not exist" containerID="ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b" Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.077261 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b"} err="failed to get container status \"ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b\": rpc error: code = NotFound desc = could not find container \"ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b\": container with ID starting with ec4212db894822e25c7894172c88884e717a445af95dac1d95efa74c4d051d9b not found: ID does not exist" Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.077291 4715 scope.go:117] "RemoveContainer" containerID="f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e" Dec 03 22:17:31 crc kubenswrapper[4715]: E1203 22:17:31.077721 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e\": container with ID starting with f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e not found: ID does not exist" containerID="f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e" Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.077750 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e"} err="failed to get container status \"f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e\": rpc error: code = NotFound desc = could not find container \"f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e\": container with ID starting with f541831fc47e15b148ada6968613df4ca29d0f3533c5c523a1969fc3b00e4f8e not found: ID does not exist" Dec 03 22:17:31 crc kubenswrapper[4715]: I1203 22:17:31.648682 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" path="/var/lib/kubelet/pods/de1bec8a-6677-4be3-8722-5ec1bb336d6f/volumes" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.021264 4715 generic.go:334] "Generic (PLEG): container finished" podID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerID="cba38c86a3d4c38fafeabc21ade5e157bd44f053c133087223f32157a4646236" exitCode=0 Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.021342 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f9c554bf8-zrldg" event={"ID":"a7e6cd7d-b1f9-4be3-b611-4efb245838bc","Type":"ContainerDied","Data":"cba38c86a3d4c38fafeabc21ade5e157bd44f053c133087223f32157a4646236"} Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.324585 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.426695 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data\") pod \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.500650 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data" (OuterVolumeSpecName: "config-data") pod "a7e6cd7d-b1f9-4be3-b611-4efb245838bc" (UID: "a7e6cd7d-b1f9-4be3-b611-4efb245838bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.527911 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfnbq\" (UniqueName: \"kubernetes.io/projected/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-kube-api-access-tfnbq\") pod \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.528077 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-combined-ca-bundle\") pod \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.528145 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data-custom\") pod \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.528197 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-logs\") pod \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\" (UID: \"a7e6cd7d-b1f9-4be3-b611-4efb245838bc\") " Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.528709 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.528874 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-logs" (OuterVolumeSpecName: "logs") pod "a7e6cd7d-b1f9-4be3-b611-4efb245838bc" (UID: "a7e6cd7d-b1f9-4be3-b611-4efb245838bc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.532194 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-kube-api-access-tfnbq" (OuterVolumeSpecName: "kube-api-access-tfnbq") pod "a7e6cd7d-b1f9-4be3-b611-4efb245838bc" (UID: "a7e6cd7d-b1f9-4be3-b611-4efb245838bc"). InnerVolumeSpecName "kube-api-access-tfnbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.532319 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a7e6cd7d-b1f9-4be3-b611-4efb245838bc" (UID: "a7e6cd7d-b1f9-4be3-b611-4efb245838bc"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.538173 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.588708 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.604115 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7e6cd7d-b1f9-4be3-b611-4efb245838bc" (UID: "a7e6cd7d-b1f9-4be3-b611-4efb245838bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.613303 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.637876 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.637929 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.637945 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.637962 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfnbq\" (UniqueName: \"kubernetes.io/projected/a7e6cd7d-b1f9-4be3-b611-4efb245838bc-kube-api-access-tfnbq\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.689076 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-srvlb"] Dec 03 22:17:32 crc kubenswrapper[4715]: I1203 22:17:32.689368 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" podUID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerName="dnsmasq-dns" containerID="cri-o://9ac3750e10043c097ca07f6f5b8dc3fce76ad776d5df17e0caf202c6ab68e576" gracePeriod=10 Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.035552 4715 generic.go:334] "Generic (PLEG): container finished" podID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerID="9ac3750e10043c097ca07f6f5b8dc3fce76ad776d5df17e0caf202c6ab68e576" exitCode=0 Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.035627 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" event={"ID":"1812c74c-bfdd-4707-a57e-c2ae44813d19","Type":"ContainerDied","Data":"9ac3750e10043c097ca07f6f5b8dc3fce76ad776d5df17e0caf202c6ab68e576"} Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.037957 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f9c554bf8-zrldg" event={"ID":"a7e6cd7d-b1f9-4be3-b611-4efb245838bc","Type":"ContainerDied","Data":"5b8b96ca47fa76f111a51a630d594ee307c08316fc66bc49dc32c4e330cb0046"} Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.038001 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f9c554bf8-zrldg" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.038017 4715 scope.go:117] "RemoveContainer" containerID="cba38c86a3d4c38fafeabc21ade5e157bd44f053c133087223f32157a4646236" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.038154 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerName="probe" containerID="cri-o://e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61" gracePeriod=30 Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.038107 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerName="cinder-scheduler" containerID="cri-o://261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a" gracePeriod=30 Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.087071 4715 scope.go:117] "RemoveContainer" containerID="5e2f2776663015bdd94f6fe6f38549d2c131e9bf5ac3ac0193d027e67d5919d5" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.091425 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-7f9c554bf8-zrldg"] Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.097143 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-7f9c554bf8-zrldg"] Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.112040 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.249682 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-sb\") pod \"1812c74c-bfdd-4707-a57e-c2ae44813d19\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.249751 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j8ph\" (UniqueName: \"kubernetes.io/projected/1812c74c-bfdd-4707-a57e-c2ae44813d19-kube-api-access-8j8ph\") pod \"1812c74c-bfdd-4707-a57e-c2ae44813d19\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.249796 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-config\") pod \"1812c74c-bfdd-4707-a57e-c2ae44813d19\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.249937 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-nb\") pod \"1812c74c-bfdd-4707-a57e-c2ae44813d19\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.249975 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-swift-storage-0\") pod \"1812c74c-bfdd-4707-a57e-c2ae44813d19\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.250004 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-svc\") pod \"1812c74c-bfdd-4707-a57e-c2ae44813d19\" (UID: \"1812c74c-bfdd-4707-a57e-c2ae44813d19\") " Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.259484 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1812c74c-bfdd-4707-a57e-c2ae44813d19-kube-api-access-8j8ph" (OuterVolumeSpecName: "kube-api-access-8j8ph") pod "1812c74c-bfdd-4707-a57e-c2ae44813d19" (UID: "1812c74c-bfdd-4707-a57e-c2ae44813d19"). InnerVolumeSpecName "kube-api-access-8j8ph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.320838 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1812c74c-bfdd-4707-a57e-c2ae44813d19" (UID: "1812c74c-bfdd-4707-a57e-c2ae44813d19"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.321429 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1812c74c-bfdd-4707-a57e-c2ae44813d19" (UID: "1812c74c-bfdd-4707-a57e-c2ae44813d19"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.328722 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1812c74c-bfdd-4707-a57e-c2ae44813d19" (UID: "1812c74c-bfdd-4707-a57e-c2ae44813d19"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.337439 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1812c74c-bfdd-4707-a57e-c2ae44813d19" (UID: "1812c74c-bfdd-4707-a57e-c2ae44813d19"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.347803 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-config" (OuterVolumeSpecName: "config") pod "1812c74c-bfdd-4707-a57e-c2ae44813d19" (UID: "1812c74c-bfdd-4707-a57e-c2ae44813d19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.354641 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.354828 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j8ph\" (UniqueName: \"kubernetes.io/projected/1812c74c-bfdd-4707-a57e-c2ae44813d19-kube-api-access-8j8ph\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.354891 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.355005 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.355069 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.355128 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1812c74c-bfdd-4707-a57e-c2ae44813d19-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:33 crc kubenswrapper[4715]: I1203 22:17:33.663981 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" path="/var/lib/kubelet/pods/a7e6cd7d-b1f9-4be3-b611-4efb245838bc/volumes" Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.053146 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.053147 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" event={"ID":"1812c74c-bfdd-4707-a57e-c2ae44813d19","Type":"ContainerDied","Data":"416c3ed228f1152a5b24a159dc38130a969efb9a1d7af78258f25fb123b64f73"} Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.053608 4715 scope.go:117] "RemoveContainer" containerID="9ac3750e10043c097ca07f6f5b8dc3fce76ad776d5df17e0caf202c6ab68e576" Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.059942 4715 generic.go:334] "Generic (PLEG): container finished" podID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerID="e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61" exitCode=0 Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.059984 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"905c7551-4b61-4cbe-a440-9a786e863ed1","Type":"ContainerDied","Data":"e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61"} Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.078454 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-srvlb"] Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.086118 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-srvlb"] Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.087656 4715 scope.go:117] "RemoveContainer" containerID="f7b2eead7bfb1f642128f6458ee351ce9acd51f97d015fa7e884738fd8f031e1" Dec 03 22:17:34 crc kubenswrapper[4715]: I1203 22:17:34.343875 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-76bcbf7d4d-l7qmc" Dec 03 22:17:35 crc kubenswrapper[4715]: I1203 22:17:35.649594 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1812c74c-bfdd-4707-a57e-c2ae44813d19" path="/var/lib/kubelet/pods/1812c74c-bfdd-4707-a57e-c2ae44813d19/volumes" Dec 03 22:17:37 crc kubenswrapper[4715]: I1203 22:17:37.936136 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.008317 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-848cf88cfc-srvlb" podUID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.157:5353: i/o timeout" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.051863 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-combined-ca-bundle\") pod \"905c7551-4b61-4cbe-a440-9a786e863ed1\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.052171 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data-custom\") pod \"905c7551-4b61-4cbe-a440-9a786e863ed1\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.052288 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn6nz\" (UniqueName: \"kubernetes.io/projected/905c7551-4b61-4cbe-a440-9a786e863ed1-kube-api-access-cn6nz\") pod \"905c7551-4b61-4cbe-a440-9a786e863ed1\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.052409 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data\") pod \"905c7551-4b61-4cbe-a440-9a786e863ed1\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.052467 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-scripts\") pod \"905c7551-4b61-4cbe-a440-9a786e863ed1\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.052581 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/905c7551-4b61-4cbe-a440-9a786e863ed1-etc-machine-id\") pod \"905c7551-4b61-4cbe-a440-9a786e863ed1\" (UID: \"905c7551-4b61-4cbe-a440-9a786e863ed1\") " Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.053168 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/905c7551-4b61-4cbe-a440-9a786e863ed1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "905c7551-4b61-4cbe-a440-9a786e863ed1" (UID: "905c7551-4b61-4cbe-a440-9a786e863ed1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.061614 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "905c7551-4b61-4cbe-a440-9a786e863ed1" (UID: "905c7551-4b61-4cbe-a440-9a786e863ed1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.061663 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-scripts" (OuterVolumeSpecName: "scripts") pod "905c7551-4b61-4cbe-a440-9a786e863ed1" (UID: "905c7551-4b61-4cbe-a440-9a786e863ed1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.061681 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/905c7551-4b61-4cbe-a440-9a786e863ed1-kube-api-access-cn6nz" (OuterVolumeSpecName: "kube-api-access-cn6nz") pod "905c7551-4b61-4cbe-a440-9a786e863ed1" (UID: "905c7551-4b61-4cbe-a440-9a786e863ed1"). InnerVolumeSpecName "kube-api-access-cn6nz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.102285 4715 generic.go:334] "Generic (PLEG): container finished" podID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerID="261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a" exitCode=0 Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.102322 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"905c7551-4b61-4cbe-a440-9a786e863ed1","Type":"ContainerDied","Data":"261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a"} Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.102348 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"905c7551-4b61-4cbe-a440-9a786e863ed1","Type":"ContainerDied","Data":"10df1a48a9a81a3391c7d9f70df1a00a961aaa2a94a4cecdb7532784e58af744"} Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.102364 4715 scope.go:117] "RemoveContainer" containerID="e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.102482 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.131761 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "905c7551-4b61-4cbe-a440-9a786e863ed1" (UID: "905c7551-4b61-4cbe-a440-9a786e863ed1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.156476 4715 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/905c7551-4b61-4cbe-a440-9a786e863ed1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.156672 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.156699 4715 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.156714 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn6nz\" (UniqueName: \"kubernetes.io/projected/905c7551-4b61-4cbe-a440-9a786e863ed1-kube-api-access-cn6nz\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.156756 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.162610 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data" (OuterVolumeSpecName: "config-data") pod "905c7551-4b61-4cbe-a440-9a786e863ed1" (UID: "905c7551-4b61-4cbe-a440-9a786e863ed1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.179144 4715 scope.go:117] "RemoveContainer" containerID="261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.196645 4715 scope.go:117] "RemoveContainer" containerID="e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.197085 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61\": container with ID starting with e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61 not found: ID does not exist" containerID="e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.197129 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61"} err="failed to get container status \"e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61\": rpc error: code = NotFound desc = could not find container \"e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61\": container with ID starting with e2a5bae6d2badb7ea9e661243d2ac370db87eea1b22a4f1c9d682896a4ad4c61 not found: ID does not exist" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.197154 4715 scope.go:117] "RemoveContainer" containerID="261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.197482 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a\": container with ID starting with 261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a not found: ID does not exist" containerID="261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.197517 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a"} err="failed to get container status \"261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a\": rpc error: code = NotFound desc = could not find container \"261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a\": container with ID starting with 261decd8cf07c8f85aba9cd3af58792e3c9be7c4bb4debc026f79bcf0b254a6a not found: ID does not exist" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.258879 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/905c7551-4b61-4cbe-a440-9a786e863ed1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.314625 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-74cd84ccf4-k9bcn" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.440358 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.449890 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.462648 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.462971 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerName="dnsmasq-dns" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.462987 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerName="dnsmasq-dns" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.463004 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerName="neutron-httpd" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463011 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerName="neutron-httpd" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.463018 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api-log" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463024 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api-log" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.463033 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463039 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.463051 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerName="probe" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463059 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerName="probe" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.463066 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerName="neutron-api" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463072 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerName="neutron-api" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.463092 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerName="cinder-scheduler" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463097 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerName="cinder-scheduler" Dec 03 22:17:38 crc kubenswrapper[4715]: E1203 22:17:38.463113 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerName="init" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463118 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerName="init" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463276 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerName="cinder-scheduler" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463288 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1812c74c-bfdd-4707-a57e-c2ae44813d19" containerName="dnsmasq-dns" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463297 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerName="neutron-api" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463305 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api-log" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463319 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" containerName="probe" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463329 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7e6cd7d-b1f9-4be3-b611-4efb245838bc" containerName="barbican-api" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.463338 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="de1bec8a-6677-4be3-8722-5ec1bb336d6f" containerName="neutron-httpd" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.464220 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.466915 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.476455 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.564366 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.564417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.564448 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.564514 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-config-data\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.564551 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqhmn\" (UniqueName: \"kubernetes.io/projected/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-kube-api-access-nqhmn\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.564597 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-scripts\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.669282 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.669335 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.669368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.669431 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-config-data\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.669482 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqhmn\" (UniqueName: \"kubernetes.io/projected/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-kube-api-access-nqhmn\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.669581 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-scripts\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.670847 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.675906 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-scripts\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.677244 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.678315 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.683897 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-config-data\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.695064 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqhmn\" (UniqueName: \"kubernetes.io/projected/cad3e724-7c3e-498d-9a7a-362a4f57c2b5-kube-api-access-nqhmn\") pod \"cinder-scheduler-0\" (UID: \"cad3e724-7c3e-498d-9a7a-362a4f57c2b5\") " pod="openstack/cinder-scheduler-0" Dec 03 22:17:38 crc kubenswrapper[4715]: I1203 22:17:38.784330 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.179748 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.198643 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.285594 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.287227 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.291296 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.291460 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.291852 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-fvlzf" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.297870 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.490120 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/205269ef-3624-4663-957b-907cf1a7cadc-openstack-config\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.490243 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/205269ef-3624-4663-957b-907cf1a7cadc-openstack-config-secret\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.490352 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205269ef-3624-4663-957b-907cf1a7cadc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.490405 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9t9d\" (UniqueName: \"kubernetes.io/projected/205269ef-3624-4663-957b-907cf1a7cadc-kube-api-access-k9t9d\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.591294 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/205269ef-3624-4663-957b-907cf1a7cadc-openstack-config\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.591360 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/205269ef-3624-4663-957b-907cf1a7cadc-openstack-config-secret\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.591416 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205269ef-3624-4663-957b-907cf1a7cadc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.591442 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9t9d\" (UniqueName: \"kubernetes.io/projected/205269ef-3624-4663-957b-907cf1a7cadc-kube-api-access-k9t9d\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.592852 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/205269ef-3624-4663-957b-907cf1a7cadc-openstack-config\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.600881 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/205269ef-3624-4663-957b-907cf1a7cadc-combined-ca-bundle\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.601010 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/205269ef-3624-4663-957b-907cf1a7cadc-openstack-config-secret\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.609195 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9t9d\" (UniqueName: \"kubernetes.io/projected/205269ef-3624-4663-957b-907cf1a7cadc-kube-api-access-k9t9d\") pod \"openstackclient\" (UID: \"205269ef-3624-4663-957b-907cf1a7cadc\") " pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.621218 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 22:17:39 crc kubenswrapper[4715]: I1203 22:17:39.655196 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="905c7551-4b61-4cbe-a440-9a786e863ed1" path="/var/lib/kubelet/pods/905c7551-4b61-4cbe-a440-9a786e863ed1/volumes" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.131541 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cad3e724-7c3e-498d-9a7a-362a4f57c2b5","Type":"ContainerStarted","Data":"0f40081abafb7ed491181f5e65167d8a5bb113db0dc9bbd3d57667e9ab2a3f7f"} Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.131789 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cad3e724-7c3e-498d-9a7a-362a4f57c2b5","Type":"ContainerStarted","Data":"0cd1e7a9ac673920fc6d0aee715778bae1921f915563da5e8b6f407427431752"} Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.228843 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 22:17:40 crc kubenswrapper[4715]: W1203 22:17:40.239519 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod205269ef_3624_4663_957b_907cf1a7cadc.slice/crio-6750e0d7c5072c35d10edc0591625631d49191ca43319d9c5fb5f103b149df7f WatchSource:0}: Error finding container 6750e0d7c5072c35d10edc0591625631d49191ca43319d9c5fb5f103b149df7f: Status 404 returned error can't find the container with id 6750e0d7c5072c35d10edc0591625631d49191ca43319d9c5fb5f103b149df7f Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.249263 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5df8bbbf89-2tsr9"] Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.250857 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.253642 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.254200 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.254515 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.265431 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5df8bbbf89-2tsr9"] Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.407976 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bf3ff4d-5720-4151-ad53-3683f08c21c5-run-httpd\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.408283 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b6zt\" (UniqueName: \"kubernetes.io/projected/1bf3ff4d-5720-4151-ad53-3683f08c21c5-kube-api-access-5b6zt\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.408346 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-internal-tls-certs\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.408382 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bf3ff4d-5720-4151-ad53-3683f08c21c5-log-httpd\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.408529 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1bf3ff4d-5720-4151-ad53-3683f08c21c5-etc-swift\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.408566 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-combined-ca-bundle\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.408641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-public-tls-certs\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.408663 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-config-data\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.509908 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bf3ff4d-5720-4151-ad53-3683f08c21c5-log-httpd\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.509944 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1bf3ff4d-5720-4151-ad53-3683f08c21c5-etc-swift\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.509964 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-combined-ca-bundle\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.510003 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-public-tls-certs\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.510020 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-config-data\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.510815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bf3ff4d-5720-4151-ad53-3683f08c21c5-log-httpd\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.511024 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bf3ff4d-5720-4151-ad53-3683f08c21c5-run-httpd\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.511044 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b6zt\" (UniqueName: \"kubernetes.io/projected/1bf3ff4d-5720-4151-ad53-3683f08c21c5-kube-api-access-5b6zt\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.511077 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-internal-tls-certs\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.511393 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bf3ff4d-5720-4151-ad53-3683f08c21c5-run-httpd\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.516068 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-combined-ca-bundle\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.519131 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-public-tls-certs\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.519695 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-internal-tls-certs\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.519912 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/1bf3ff4d-5720-4151-ad53-3683f08c21c5-etc-swift\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.520251 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bf3ff4d-5720-4151-ad53-3683f08c21c5-config-data\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.534211 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b6zt\" (UniqueName: \"kubernetes.io/projected/1bf3ff4d-5720-4151-ad53-3683f08c21c5-kube-api-access-5b6zt\") pod \"swift-proxy-5df8bbbf89-2tsr9\" (UID: \"1bf3ff4d-5720-4151-ad53-3683f08c21c5\") " pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.583184 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.748921 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.749403 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="ceilometer-central-agent" containerID="cri-o://ca95a4cb86b25c3c32bc5f98fcc82f447b929f396069f8e09590cfe8a5a40ceb" gracePeriod=30 Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.749555 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="proxy-httpd" containerID="cri-o://fa615843941ce2dc9968739005e78812be531b42166aabb607eb92cb2f231dcc" gracePeriod=30 Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.749571 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="ceilometer-notification-agent" containerID="cri-o://128f27f9955d3da7ff215a33f665159a95a15070dbc5c6de235f1db3c70e9993" gracePeriod=30 Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.749769 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="sg-core" containerID="cri-o://71132a3bc090ff4256672c9f66529c3304afd4b9efb8fc7179452076ef42610e" gracePeriod=30 Dec 03 22:17:40 crc kubenswrapper[4715]: I1203 22:17:40.853166 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.159:3000/\": read tcp 10.217.0.2:60126->10.217.0.159:3000: read: connection reset by peer" Dec 03 22:17:41 crc kubenswrapper[4715]: I1203 22:17:41.142027 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"205269ef-3624-4663-957b-907cf1a7cadc","Type":"ContainerStarted","Data":"6750e0d7c5072c35d10edc0591625631d49191ca43319d9c5fb5f103b149df7f"} Dec 03 22:17:41 crc kubenswrapper[4715]: I1203 22:17:41.161748 4715 generic.go:334] "Generic (PLEG): container finished" podID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerID="fa615843941ce2dc9968739005e78812be531b42166aabb607eb92cb2f231dcc" exitCode=0 Dec 03 22:17:41 crc kubenswrapper[4715]: I1203 22:17:41.161780 4715 generic.go:334] "Generic (PLEG): container finished" podID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerID="71132a3bc090ff4256672c9f66529c3304afd4b9efb8fc7179452076ef42610e" exitCode=2 Dec 03 22:17:41 crc kubenswrapper[4715]: I1203 22:17:41.161824 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerDied","Data":"fa615843941ce2dc9968739005e78812be531b42166aabb607eb92cb2f231dcc"} Dec 03 22:17:41 crc kubenswrapper[4715]: I1203 22:17:41.161850 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerDied","Data":"71132a3bc090ff4256672c9f66529c3304afd4b9efb8fc7179452076ef42610e"} Dec 03 22:17:41 crc kubenswrapper[4715]: I1203 22:17:41.163567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"cad3e724-7c3e-498d-9a7a-362a4f57c2b5","Type":"ContainerStarted","Data":"95fed11a73f1e6f9fb2008e60149a920c897e05a077edc0e86d67e047f8270be"} Dec 03 22:17:41 crc kubenswrapper[4715]: I1203 22:17:41.173317 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5df8bbbf89-2tsr9"] Dec 03 22:17:41 crc kubenswrapper[4715]: I1203 22:17:41.190564 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.190546086 podStartE2EDuration="3.190546086s" podCreationTimestamp="2025-12-03 22:17:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:41.184029101 +0000 UTC m=+2157.926739696" watchObservedRunningTime="2025-12-03 22:17:41.190546086 +0000 UTC m=+2157.933256681" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.181311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" event={"ID":"1bf3ff4d-5720-4151-ad53-3683f08c21c5","Type":"ContainerStarted","Data":"e77a9b60f30b7ca4448613964391530a71980480d12c74cb2a72fac9dccc517a"} Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.181884 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.181898 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" event={"ID":"1bf3ff4d-5720-4151-ad53-3683f08c21c5","Type":"ContainerStarted","Data":"b1d33688a5fc9eaae4400a0b84c7dc03c18847073db8c3cb3e952b9127925ac8"} Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.181910 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" event={"ID":"1bf3ff4d-5720-4151-ad53-3683f08c21c5","Type":"ContainerStarted","Data":"0e736777a834c6f3155b836da5a8853370516ff2d10b57cd2b2c451533c054ff"} Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.186265 4715 generic.go:334] "Generic (PLEG): container finished" podID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerID="128f27f9955d3da7ff215a33f665159a95a15070dbc5c6de235f1db3c70e9993" exitCode=0 Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.186309 4715 generic.go:334] "Generic (PLEG): container finished" podID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerID="ca95a4cb86b25c3c32bc5f98fcc82f447b929f396069f8e09590cfe8a5a40ceb" exitCode=0 Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.186397 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerDied","Data":"128f27f9955d3da7ff215a33f665159a95a15070dbc5c6de235f1db3c70e9993"} Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.186423 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerDied","Data":"ca95a4cb86b25c3c32bc5f98fcc82f447b929f396069f8e09590cfe8a5a40ceb"} Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.212985 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" podStartSLOduration=2.212969468 podStartE2EDuration="2.212969468s" podCreationTimestamp="2025-12-03 22:17:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:17:42.205029945 +0000 UTC m=+2158.947740540" watchObservedRunningTime="2025-12-03 22:17:42.212969468 +0000 UTC m=+2158.955680063" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.570328 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.655924 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-combined-ca-bundle\") pod \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.656240 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-log-httpd\") pod \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.656313 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-scripts\") pod \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.656453 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-run-httpd\") pod \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.656472 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-sg-core-conf-yaml\") pod \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.656545 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-config-data\") pod \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.656599 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h65nt\" (UniqueName: \"kubernetes.io/projected/3b11ee3d-7aea-416f-b035-1ad5efe949e2-kube-api-access-h65nt\") pod \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\" (UID: \"3b11ee3d-7aea-416f-b035-1ad5efe949e2\") " Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.656721 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3b11ee3d-7aea-416f-b035-1ad5efe949e2" (UID: "3b11ee3d-7aea-416f-b035-1ad5efe949e2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.656793 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3b11ee3d-7aea-416f-b035-1ad5efe949e2" (UID: "3b11ee3d-7aea-416f-b035-1ad5efe949e2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.657513 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.657533 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3b11ee3d-7aea-416f-b035-1ad5efe949e2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.663619 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-scripts" (OuterVolumeSpecName: "scripts") pod "3b11ee3d-7aea-416f-b035-1ad5efe949e2" (UID: "3b11ee3d-7aea-416f-b035-1ad5efe949e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.664627 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b11ee3d-7aea-416f-b035-1ad5efe949e2-kube-api-access-h65nt" (OuterVolumeSpecName: "kube-api-access-h65nt") pod "3b11ee3d-7aea-416f-b035-1ad5efe949e2" (UID: "3b11ee3d-7aea-416f-b035-1ad5efe949e2"). InnerVolumeSpecName "kube-api-access-h65nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.686340 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3b11ee3d-7aea-416f-b035-1ad5efe949e2" (UID: "3b11ee3d-7aea-416f-b035-1ad5efe949e2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.760409 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h65nt\" (UniqueName: \"kubernetes.io/projected/3b11ee3d-7aea-416f-b035-1ad5efe949e2-kube-api-access-h65nt\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.760437 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.760446 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.762640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-config-data" (OuterVolumeSpecName: "config-data") pod "3b11ee3d-7aea-416f-b035-1ad5efe949e2" (UID: "3b11ee3d-7aea-416f-b035-1ad5efe949e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.779833 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b11ee3d-7aea-416f-b035-1ad5efe949e2" (UID: "3b11ee3d-7aea-416f-b035-1ad5efe949e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.862103 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:42 crc kubenswrapper[4715]: I1203 22:17:42.862318 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b11ee3d-7aea-416f-b035-1ad5efe949e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.200721 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3b11ee3d-7aea-416f-b035-1ad5efe949e2","Type":"ContainerDied","Data":"c4b44f0a3896c1c5685c7cc46c1dadbaa949c81720dbbc07c711fb9029c181e1"} Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.200784 4715 scope.go:117] "RemoveContainer" containerID="fa615843941ce2dc9968739005e78812be531b42166aabb607eb92cb2f231dcc" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.200913 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.201006 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.223316 4715 scope.go:117] "RemoveContainer" containerID="71132a3bc090ff4256672c9f66529c3304afd4b9efb8fc7179452076ef42610e" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.247963 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.257315 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.270418 4715 scope.go:117] "RemoveContainer" containerID="128f27f9955d3da7ff215a33f665159a95a15070dbc5c6de235f1db3c70e9993" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.270756 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:43 crc kubenswrapper[4715]: E1203 22:17:43.271137 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="sg-core" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.271156 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="sg-core" Dec 03 22:17:43 crc kubenswrapper[4715]: E1203 22:17:43.271176 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="proxy-httpd" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.271182 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="proxy-httpd" Dec 03 22:17:43 crc kubenswrapper[4715]: E1203 22:17:43.271194 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="ceilometer-central-agent" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.271202 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="ceilometer-central-agent" Dec 03 22:17:43 crc kubenswrapper[4715]: E1203 22:17:43.271214 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="ceilometer-notification-agent" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.271220 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="ceilometer-notification-agent" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.271418 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="ceilometer-central-agent" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.271436 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="proxy-httpd" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.271452 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="sg-core" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.271467 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" containerName="ceilometer-notification-agent" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.273474 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.277321 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.279060 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.284984 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.313632 4715 scope.go:117] "RemoveContainer" containerID="ca95a4cb86b25c3c32bc5f98fcc82f447b929f396069f8e09590cfe8a5a40ceb" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.372549 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-config-data\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.372636 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrjml\" (UniqueName: \"kubernetes.io/projected/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-kube-api-access-wrjml\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.372734 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.372759 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.372798 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-log-httpd\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.372838 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-run-httpd\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.372910 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-scripts\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.475089 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.475136 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.475165 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-log-httpd\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.475207 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-run-httpd\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.475234 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-scripts\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.475341 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-config-data\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.475360 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrjml\" (UniqueName: \"kubernetes.io/projected/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-kube-api-access-wrjml\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.476134 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-run-httpd\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.476237 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-log-httpd\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.479730 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-scripts\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.488142 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.488147 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.488673 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-config-data\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.497566 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrjml\" (UniqueName: \"kubernetes.io/projected/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-kube-api-access-wrjml\") pod \"ceilometer-0\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.603326 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.658672 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b11ee3d-7aea-416f-b035-1ad5efe949e2" path="/var/lib/kubelet/pods/3b11ee3d-7aea-416f-b035-1ad5efe949e2/volumes" Dec 03 22:17:43 crc kubenswrapper[4715]: I1203 22:17:43.785868 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 22:17:44 crc kubenswrapper[4715]: I1203 22:17:44.160904 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:44 crc kubenswrapper[4715]: I1203 22:17:44.212858 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerStarted","Data":"0536f82c9566454f44b4f1f785e3a486dc81b5bf5f77151e3367c593838e6000"} Dec 03 22:17:45 crc kubenswrapper[4715]: I1203 22:17:45.227328 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerStarted","Data":"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936"} Dec 03 22:17:48 crc kubenswrapper[4715]: I1203 22:17:48.315680 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-74cd84ccf4-k9bcn" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 03 22:17:49 crc kubenswrapper[4715]: I1203 22:17:49.042094 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 22:17:49 crc kubenswrapper[4715]: I1203 22:17:49.173361 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:50 crc kubenswrapper[4715]: I1203 22:17:50.591420 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:50 crc kubenswrapper[4715]: I1203 22:17:50.601477 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5df8bbbf89-2tsr9" Dec 03 22:17:51 crc kubenswrapper[4715]: I1203 22:17:51.280745 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"205269ef-3624-4663-957b-907cf1a7cadc","Type":"ContainerStarted","Data":"5638bbc0f40803691f678846b007890d26e035c64a58c139d6f7a42a70313f39"} Dec 03 22:17:51 crc kubenswrapper[4715]: I1203 22:17:51.285226 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerStarted","Data":"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e"} Dec 03 22:17:52 crc kubenswrapper[4715]: I1203 22:17:52.322981 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.713770418 podStartE2EDuration="13.322965326s" podCreationTimestamp="2025-12-03 22:17:39 +0000 UTC" firstStartedPulling="2025-12-03 22:17:40.242199044 +0000 UTC m=+2156.984909639" lastFinishedPulling="2025-12-03 22:17:50.851393942 +0000 UTC m=+2167.594104547" observedRunningTime="2025-12-03 22:17:52.317415527 +0000 UTC m=+2169.060126122" watchObservedRunningTime="2025-12-03 22:17:52.322965326 +0000 UTC m=+2169.065675921" Dec 03 22:17:53 crc kubenswrapper[4715]: E1203 22:17:53.021266 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Dec 03 22:17:53 crc kubenswrapper[4715]: I1203 22:17:53.306439 4715 generic.go:334] "Generic (PLEG): container finished" podID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerID="706dc959898b0c6a666acb6b020a24e81832e5be7099c366f91458474d2b0618" exitCode=137 Dec 03 22:17:53 crc kubenswrapper[4715]: I1203 22:17:53.306577 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cd84ccf4-k9bcn" event={"ID":"ae09bc5f-d3e1-48c3-8145-611fd9827870","Type":"ContainerDied","Data":"706dc959898b0c6a666acb6b020a24e81832e5be7099c366f91458474d2b0618"} Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.128513 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.306165 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-config-data\") pod \"ae09bc5f-d3e1-48c3-8145-611fd9827870\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.306280 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-tls-certs\") pod \"ae09bc5f-d3e1-48c3-8145-611fd9827870\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.306346 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sp89\" (UniqueName: \"kubernetes.io/projected/ae09bc5f-d3e1-48c3-8145-611fd9827870-kube-api-access-5sp89\") pod \"ae09bc5f-d3e1-48c3-8145-611fd9827870\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.306423 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-secret-key\") pod \"ae09bc5f-d3e1-48c3-8145-611fd9827870\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.306469 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae09bc5f-d3e1-48c3-8145-611fd9827870-logs\") pod \"ae09bc5f-d3e1-48c3-8145-611fd9827870\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.306647 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-combined-ca-bundle\") pod \"ae09bc5f-d3e1-48c3-8145-611fd9827870\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.306713 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-scripts\") pod \"ae09bc5f-d3e1-48c3-8145-611fd9827870\" (UID: \"ae09bc5f-d3e1-48c3-8145-611fd9827870\") " Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.307128 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae09bc5f-d3e1-48c3-8145-611fd9827870-logs" (OuterVolumeSpecName: "logs") pod "ae09bc5f-d3e1-48c3-8145-611fd9827870" (UID: "ae09bc5f-d3e1-48c3-8145-611fd9827870"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.307311 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae09bc5f-d3e1-48c3-8145-611fd9827870-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.311243 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ae09bc5f-d3e1-48c3-8145-611fd9827870" (UID: "ae09bc5f-d3e1-48c3-8145-611fd9827870"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.313034 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae09bc5f-d3e1-48c3-8145-611fd9827870-kube-api-access-5sp89" (OuterVolumeSpecName: "kube-api-access-5sp89") pod "ae09bc5f-d3e1-48c3-8145-611fd9827870" (UID: "ae09bc5f-d3e1-48c3-8145-611fd9827870"). InnerVolumeSpecName "kube-api-access-5sp89". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.322437 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74cd84ccf4-k9bcn" event={"ID":"ae09bc5f-d3e1-48c3-8145-611fd9827870","Type":"ContainerDied","Data":"2e99021e5e3860269e29c5cca98f3ef145efad242585dad019ddf055dff63714"} Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.322487 4715 scope.go:117] "RemoveContainer" containerID="4979bdb653c8db320223f4fb81a9d23bee0c19d8f6be47165c7af4d53c639d1f" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.322638 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74cd84ccf4-k9bcn" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.326989 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerStarted","Data":"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a"} Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.348664 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae09bc5f-d3e1-48c3-8145-611fd9827870" (UID: "ae09bc5f-d3e1-48c3-8145-611fd9827870"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.349160 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-scripts" (OuterVolumeSpecName: "scripts") pod "ae09bc5f-d3e1-48c3-8145-611fd9827870" (UID: "ae09bc5f-d3e1-48c3-8145-611fd9827870"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.365309 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-config-data" (OuterVolumeSpecName: "config-data") pod "ae09bc5f-d3e1-48c3-8145-611fd9827870" (UID: "ae09bc5f-d3e1-48c3-8145-611fd9827870"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.368451 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "ae09bc5f-d3e1-48c3-8145-611fd9827870" (UID: "ae09bc5f-d3e1-48c3-8145-611fd9827870"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.409302 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.409333 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sp89\" (UniqueName: \"kubernetes.io/projected/ae09bc5f-d3e1-48c3-8145-611fd9827870-kube-api-access-5sp89\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.409344 4715 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.409352 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae09bc5f-d3e1-48c3-8145-611fd9827870-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.409361 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.409372 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ae09bc5f-d3e1-48c3-8145-611fd9827870-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.519172 4715 scope.go:117] "RemoveContainer" containerID="706dc959898b0c6a666acb6b020a24e81832e5be7099c366f91458474d2b0618" Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.656136 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-74cd84ccf4-k9bcn"] Dec 03 22:17:54 crc kubenswrapper[4715]: I1203 22:17:54.662938 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-74cd84ccf4-k9bcn"] Dec 03 22:17:55 crc kubenswrapper[4715]: I1203 22:17:55.649437 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" path="/var/lib/kubelet/pods/ae09bc5f-d3e1-48c3-8145-611fd9827870/volumes" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.204003 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pdwx4"] Dec 03 22:17:56 crc kubenswrapper[4715]: E1203 22:17:56.204457 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.204479 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" Dec 03 22:17:56 crc kubenswrapper[4715]: E1203 22:17:56.204514 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon-log" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.204523 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon-log" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.204758 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.204784 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae09bc5f-d3e1-48c3-8145-611fd9827870" containerName="horizon-log" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.206426 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.215397 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pdwx4"] Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.347967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerStarted","Data":"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51"} Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.348127 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="ceilometer-central-agent" containerID="cri-o://4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936" gracePeriod=30 Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.348158 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.348199 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="sg-core" containerID="cri-o://10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a" gracePeriod=30 Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.348200 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="proxy-httpd" containerID="cri-o://2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51" gracePeriod=30 Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.348227 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="ceilometer-notification-agent" containerID="cri-o://bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e" gracePeriod=30 Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.364823 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-utilities\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.365592 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-catalog-content\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.366002 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvdj6\" (UniqueName: \"kubernetes.io/projected/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-kube-api-access-zvdj6\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.380348 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.024990489 podStartE2EDuration="13.380331339s" podCreationTimestamp="2025-12-03 22:17:43 +0000 UTC" firstStartedPulling="2025-12-03 22:17:44.176855918 +0000 UTC m=+2160.919566523" lastFinishedPulling="2025-12-03 22:17:55.532196768 +0000 UTC m=+2172.274907373" observedRunningTime="2025-12-03 22:17:56.374696438 +0000 UTC m=+2173.117407033" watchObservedRunningTime="2025-12-03 22:17:56.380331339 +0000 UTC m=+2173.123041934" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.468023 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-utilities\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.468137 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-catalog-content\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.468188 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvdj6\" (UniqueName: \"kubernetes.io/projected/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-kube-api-access-zvdj6\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.468459 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-utilities\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.468636 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-catalog-content\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.485858 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvdj6\" (UniqueName: \"kubernetes.io/projected/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-kube-api-access-zvdj6\") pod \"certified-operators-pdwx4\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:56 crc kubenswrapper[4715]: I1203 22:17:56.565322 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:17:57 crc kubenswrapper[4715]: W1203 22:17:57.055642 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9fe015d5_6916_4c9a_a20f_d8be799a0e4c.slice/crio-a167ef45883d0b839ac5a6f660c1162250a2383f6f9118e83ebbbf55c1acfe97 WatchSource:0}: Error finding container a167ef45883d0b839ac5a6f660c1162250a2383f6f9118e83ebbbf55c1acfe97: Status 404 returned error can't find the container with id a167ef45883d0b839ac5a6f660c1162250a2383f6f9118e83ebbbf55c1acfe97 Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.056217 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pdwx4"] Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.119395 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.281817 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrjml\" (UniqueName: \"kubernetes.io/projected/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-kube-api-access-wrjml\") pod \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.281935 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-run-httpd\") pod \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.281997 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-scripts\") pod \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.282170 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-log-httpd\") pod \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.282229 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-combined-ca-bundle\") pod \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.282261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" (UID: "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.282274 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-config-data\") pod \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.282330 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-sg-core-conf-yaml\") pod \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\" (UID: \"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7\") " Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.282619 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" (UID: "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.282813 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.282833 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.293680 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-scripts" (OuterVolumeSpecName: "scripts") pod "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" (UID: "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.293757 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-kube-api-access-wrjml" (OuterVolumeSpecName: "kube-api-access-wrjml") pod "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" (UID: "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7"). InnerVolumeSpecName "kube-api-access-wrjml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.321333 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" (UID: "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.357728 4715 generic.go:334] "Generic (PLEG): container finished" podID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerID="0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1" exitCode=0 Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.357784 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdwx4" event={"ID":"9fe015d5-6916-4c9a-a20f-d8be799a0e4c","Type":"ContainerDied","Data":"0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1"} Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.358327 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdwx4" event={"ID":"9fe015d5-6916-4c9a-a20f-d8be799a0e4c","Type":"ContainerStarted","Data":"a167ef45883d0b839ac5a6f660c1162250a2383f6f9118e83ebbbf55c1acfe97"} Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361669 4715 generic.go:334] "Generic (PLEG): container finished" podID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerID="2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51" exitCode=0 Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361691 4715 generic.go:334] "Generic (PLEG): container finished" podID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerID="10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a" exitCode=2 Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361699 4715 generic.go:334] "Generic (PLEG): container finished" podID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerID="bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e" exitCode=0 Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361730 4715 generic.go:334] "Generic (PLEG): container finished" podID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerID="4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936" exitCode=0 Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361749 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerDied","Data":"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51"} Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361772 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerDied","Data":"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a"} Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361782 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerDied","Data":"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e"} Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361794 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerDied","Data":"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936"} Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361805 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0cf1fd55-5e7a-45e2-a35c-05953aecc7f7","Type":"ContainerDied","Data":"0536f82c9566454f44b4f1f785e3a486dc81b5bf5f77151e3367c593838e6000"} Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361826 4715 scope.go:117] "RemoveContainer" containerID="2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.361855 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.384377 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.384472 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.384595 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrjml\" (UniqueName: \"kubernetes.io/projected/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-kube-api-access-wrjml\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.389708 4715 scope.go:117] "RemoveContainer" containerID="10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.394531 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-config-data" (OuterVolumeSpecName: "config-data") pod "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" (UID: "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.396302 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" (UID: "0cf1fd55-5e7a-45e2-a35c-05953aecc7f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.414180 4715 scope.go:117] "RemoveContainer" containerID="bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.434023 4715 scope.go:117] "RemoveContainer" containerID="4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.456529 4715 scope.go:117] "RemoveContainer" containerID="2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51" Dec 03 22:17:57 crc kubenswrapper[4715]: E1203 22:17:57.467912 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": container with ID starting with 2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51 not found: ID does not exist" containerID="2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.467954 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51"} err="failed to get container status \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": rpc error: code = NotFound desc = could not find container \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": container with ID starting with 2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51 not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.467981 4715 scope.go:117] "RemoveContainer" containerID="10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a" Dec 03 22:17:57 crc kubenswrapper[4715]: E1203 22:17:57.468344 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": container with ID starting with 10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a not found: ID does not exist" containerID="10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.468392 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a"} err="failed to get container status \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": rpc error: code = NotFound desc = could not find container \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": container with ID starting with 10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.468427 4715 scope.go:117] "RemoveContainer" containerID="bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e" Dec 03 22:17:57 crc kubenswrapper[4715]: E1203 22:17:57.468779 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": container with ID starting with bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e not found: ID does not exist" containerID="bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.468804 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e"} err="failed to get container status \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": rpc error: code = NotFound desc = could not find container \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": container with ID starting with bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.468818 4715 scope.go:117] "RemoveContainer" containerID="4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936" Dec 03 22:17:57 crc kubenswrapper[4715]: E1203 22:17:57.469036 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": container with ID starting with 4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936 not found: ID does not exist" containerID="4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.469058 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936"} err="failed to get container status \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": rpc error: code = NotFound desc = could not find container \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": container with ID starting with 4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936 not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.469077 4715 scope.go:117] "RemoveContainer" containerID="2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.469474 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51"} err="failed to get container status \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": rpc error: code = NotFound desc = could not find container \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": container with ID starting with 2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51 not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.469530 4715 scope.go:117] "RemoveContainer" containerID="10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.469749 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a"} err="failed to get container status \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": rpc error: code = NotFound desc = could not find container \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": container with ID starting with 10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.469770 4715 scope.go:117] "RemoveContainer" containerID="bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.469992 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e"} err="failed to get container status \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": rpc error: code = NotFound desc = could not find container \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": container with ID starting with bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.470011 4715 scope.go:117] "RemoveContainer" containerID="4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.470258 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936"} err="failed to get container status \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": rpc error: code = NotFound desc = could not find container \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": container with ID starting with 4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936 not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.470361 4715 scope.go:117] "RemoveContainer" containerID="2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.470690 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51"} err="failed to get container status \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": rpc error: code = NotFound desc = could not find container \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": container with ID starting with 2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51 not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.470709 4715 scope.go:117] "RemoveContainer" containerID="10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.470913 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a"} err="failed to get container status \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": rpc error: code = NotFound desc = could not find container \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": container with ID starting with 10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.471001 4715 scope.go:117] "RemoveContainer" containerID="bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.471306 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e"} err="failed to get container status \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": rpc error: code = NotFound desc = could not find container \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": container with ID starting with bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.471324 4715 scope.go:117] "RemoveContainer" containerID="4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.472080 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936"} err="failed to get container status \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": rpc error: code = NotFound desc = could not find container \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": container with ID starting with 4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936 not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.472111 4715 scope.go:117] "RemoveContainer" containerID="2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.472396 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51"} err="failed to get container status \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": rpc error: code = NotFound desc = could not find container \"2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51\": container with ID starting with 2be9df8e45ea31cb0b2a2a71770e6a423ca21e54e721f8d9fa967a03728efa51 not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.472417 4715 scope.go:117] "RemoveContainer" containerID="10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.472667 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a"} err="failed to get container status \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": rpc error: code = NotFound desc = could not find container \"10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a\": container with ID starting with 10a710ca4986a0edb8a1ce7f4e0922e9affe9cf09bf2b24582ac87fe8b13ff9a not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.472686 4715 scope.go:117] "RemoveContainer" containerID="bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.472921 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e"} err="failed to get container status \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": rpc error: code = NotFound desc = could not find container \"bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e\": container with ID starting with bd27b4f517699255425de3eb85ac4a9787b20db568f4b99d85533de878453b0e not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.472939 4715 scope.go:117] "RemoveContainer" containerID="4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.473146 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936"} err="failed to get container status \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": rpc error: code = NotFound desc = could not find container \"4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936\": container with ID starting with 4521bee8effd4455dd81243124de31b91c78778118bde8f56d13db3469bb5936 not found: ID does not exist" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.486547 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.486693 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.686021 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.694338 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.714875 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:57 crc kubenswrapper[4715]: E1203 22:17:57.715250 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="ceilometer-central-agent" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.715270 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="ceilometer-central-agent" Dec 03 22:17:57 crc kubenswrapper[4715]: E1203 22:17:57.715294 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="proxy-httpd" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.715301 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="proxy-httpd" Dec 03 22:17:57 crc kubenswrapper[4715]: E1203 22:17:57.715308 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="ceilometer-notification-agent" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.715315 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="ceilometer-notification-agent" Dec 03 22:17:57 crc kubenswrapper[4715]: E1203 22:17:57.715337 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="sg-core" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.715343 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="sg-core" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.715480 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="proxy-httpd" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.715493 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="sg-core" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.715522 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="ceilometer-central-agent" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.715538 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" containerName="ceilometer-notification-agent" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.716956 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.720413 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.721745 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.743204 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.893992 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-scripts\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.894038 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-config-data\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.894092 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-run-httpd\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.894309 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.894358 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.894569 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xndf\" (UniqueName: \"kubernetes.io/projected/c851dcc7-cdd0-438f-ade3-a8d07274897b-kube-api-access-4xndf\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.894679 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-log-httpd\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.996694 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-log-httpd\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.996796 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-scripts\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.996824 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-config-data\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.996887 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-run-httpd\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.996927 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.996947 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.997014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xndf\" (UniqueName: \"kubernetes.io/projected/c851dcc7-cdd0-438f-ade3-a8d07274897b-kube-api-access-4xndf\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.997816 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-run-httpd\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:57 crc kubenswrapper[4715]: I1203 22:17:57.998424 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-log-httpd\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:58 crc kubenswrapper[4715]: I1203 22:17:58.002036 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-config-data\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:58 crc kubenswrapper[4715]: I1203 22:17:58.002058 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:58 crc kubenswrapper[4715]: I1203 22:17:58.002223 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:58 crc kubenswrapper[4715]: I1203 22:17:58.014919 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-scripts\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:58 crc kubenswrapper[4715]: I1203 22:17:58.017910 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xndf\" (UniqueName: \"kubernetes.io/projected/c851dcc7-cdd0-438f-ade3-a8d07274897b-kube-api-access-4xndf\") pod \"ceilometer-0\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " pod="openstack/ceilometer-0" Dec 03 22:17:58 crc kubenswrapper[4715]: I1203 22:17:58.041324 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:17:58 crc kubenswrapper[4715]: I1203 22:17:58.514614 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:17:58 crc kubenswrapper[4715]: W1203 22:17:58.518338 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc851dcc7_cdd0_438f_ade3_a8d07274897b.slice/crio-dc2ff9d8c95013c7f5cab1c29b0c739c7670f37fc80835fa8a25df6edc18309c WatchSource:0}: Error finding container dc2ff9d8c95013c7f5cab1c29b0c739c7670f37fc80835fa8a25df6edc18309c: Status 404 returned error can't find the container with id dc2ff9d8c95013c7f5cab1c29b0c739c7670f37fc80835fa8a25df6edc18309c Dec 03 22:17:59 crc kubenswrapper[4715]: I1203 22:17:59.387309 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdwx4" event={"ID":"9fe015d5-6916-4c9a-a20f-d8be799a0e4c","Type":"ContainerStarted","Data":"0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6"} Dec 03 22:17:59 crc kubenswrapper[4715]: I1203 22:17:59.390476 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerStarted","Data":"dc2ff9d8c95013c7f5cab1c29b0c739c7670f37fc80835fa8a25df6edc18309c"} Dec 03 22:17:59 crc kubenswrapper[4715]: I1203 22:17:59.642740 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cf1fd55-5e7a-45e2-a35c-05953aecc7f7" path="/var/lib/kubelet/pods/0cf1fd55-5e7a-45e2-a35c-05953aecc7f7/volumes" Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.155014 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.155235 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6143cec8-8a39-443c-b850-109e24ac1200" containerName="glance-log" containerID="cri-o://3c2056cd8e993c9e3b9a3c5c5e37e0849818a26f5d0eceabf1165ac94696fa2f" gracePeriod=30 Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.155368 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="6143cec8-8a39-443c-b850-109e24ac1200" containerName="glance-httpd" containerID="cri-o://dba4177ff9d25999c348f58f9bc7a868140faaff34981749ae6ecb131064a243" gracePeriod=30 Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.401910 4715 generic.go:334] "Generic (PLEG): container finished" podID="6143cec8-8a39-443c-b850-109e24ac1200" containerID="3c2056cd8e993c9e3b9a3c5c5e37e0849818a26f5d0eceabf1165ac94696fa2f" exitCode=143 Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.401997 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6143cec8-8a39-443c-b850-109e24ac1200","Type":"ContainerDied","Data":"3c2056cd8e993c9e3b9a3c5c5e37e0849818a26f5d0eceabf1165ac94696fa2f"} Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.405074 4715 generic.go:334] "Generic (PLEG): container finished" podID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerID="0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6" exitCode=0 Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.405982 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdwx4" event={"ID":"9fe015d5-6916-4c9a-a20f-d8be799a0e4c","Type":"ContainerDied","Data":"0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6"} Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.409761 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerStarted","Data":"9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8"} Dec 03 22:18:00 crc kubenswrapper[4715]: I1203 22:18:00.409801 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerStarted","Data":"539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d"} Dec 03 22:18:01 crc kubenswrapper[4715]: I1203 22:18:01.420229 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerStarted","Data":"a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0"} Dec 03 22:18:03 crc kubenswrapper[4715]: I1203 22:18:03.440939 4715 generic.go:334] "Generic (PLEG): container finished" podID="6143cec8-8a39-443c-b850-109e24ac1200" containerID="dba4177ff9d25999c348f58f9bc7a868140faaff34981749ae6ecb131064a243" exitCode=0 Dec 03 22:18:03 crc kubenswrapper[4715]: I1203 22:18:03.441282 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6143cec8-8a39-443c-b850-109e24ac1200","Type":"ContainerDied","Data":"dba4177ff9d25999c348f58f9bc7a868140faaff34981749ae6ecb131064a243"} Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.264807 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.328109 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-scripts\") pod \"6143cec8-8a39-443c-b850-109e24ac1200\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.328165 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-httpd-run\") pod \"6143cec8-8a39-443c-b850-109e24ac1200\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.328194 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-combined-ca-bundle\") pod \"6143cec8-8a39-443c-b850-109e24ac1200\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.328229 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-logs\") pod \"6143cec8-8a39-443c-b850-109e24ac1200\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.328265 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc8qp\" (UniqueName: \"kubernetes.io/projected/6143cec8-8a39-443c-b850-109e24ac1200-kube-api-access-pc8qp\") pod \"6143cec8-8a39-443c-b850-109e24ac1200\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.328293 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"6143cec8-8a39-443c-b850-109e24ac1200\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.328311 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-config-data\") pod \"6143cec8-8a39-443c-b850-109e24ac1200\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.328367 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-public-tls-certs\") pod \"6143cec8-8a39-443c-b850-109e24ac1200\" (UID: \"6143cec8-8a39-443c-b850-109e24ac1200\") " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.329572 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-logs" (OuterVolumeSpecName: "logs") pod "6143cec8-8a39-443c-b850-109e24ac1200" (UID: "6143cec8-8a39-443c-b850-109e24ac1200"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.329657 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6143cec8-8a39-443c-b850-109e24ac1200" (UID: "6143cec8-8a39-443c-b850-109e24ac1200"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.340970 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6143cec8-8a39-443c-b850-109e24ac1200-kube-api-access-pc8qp" (OuterVolumeSpecName: "kube-api-access-pc8qp") pod "6143cec8-8a39-443c-b850-109e24ac1200" (UID: "6143cec8-8a39-443c-b850-109e24ac1200"). InnerVolumeSpecName "kube-api-access-pc8qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.343718 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "6143cec8-8a39-443c-b850-109e24ac1200" (UID: "6143cec8-8a39-443c-b850-109e24ac1200"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.345989 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-scripts" (OuterVolumeSpecName: "scripts") pod "6143cec8-8a39-443c-b850-109e24ac1200" (UID: "6143cec8-8a39-443c-b850-109e24ac1200"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.368550 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6143cec8-8a39-443c-b850-109e24ac1200" (UID: "6143cec8-8a39-443c-b850-109e24ac1200"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.419479 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6143cec8-8a39-443c-b850-109e24ac1200" (UID: "6143cec8-8a39-443c-b850-109e24ac1200"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.419779 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-config-data" (OuterVolumeSpecName: "config-data") pod "6143cec8-8a39-443c-b850-109e24ac1200" (UID: "6143cec8-8a39-443c-b850-109e24ac1200"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.433259 4715 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.433297 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.433308 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.433318 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.433329 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6143cec8-8a39-443c-b850-109e24ac1200-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.433338 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc8qp\" (UniqueName: \"kubernetes.io/projected/6143cec8-8a39-443c-b850-109e24ac1200-kube-api-access-pc8qp\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.433373 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.433386 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6143cec8-8a39-443c-b850-109e24ac1200-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.449942 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"6143cec8-8a39-443c-b850-109e24ac1200","Type":"ContainerDied","Data":"9081f3e7e372b94935f2f4a428701e17b45edccd1700f7600cfcfb329b7e0215"} Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.450000 4715 scope.go:117] "RemoveContainer" containerID="dba4177ff9d25999c348f58f9bc7a868140faaff34981749ae6ecb131064a243" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.450119 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.455053 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.504404 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.504880 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="279fe539-9749-4528-b219-df6323816ea7" containerName="glance-log" containerID="cri-o://f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459" gracePeriod=30 Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.505152 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="279fe539-9749-4528-b219-df6323816ea7" containerName="glance-httpd" containerID="cri-o://a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f" gracePeriod=30 Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.535400 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.617973 4715 scope.go:117] "RemoveContainer" containerID="3c2056cd8e993c9e3b9a3c5c5e37e0849818a26f5d0eceabf1165ac94696fa2f" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.628794 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.639153 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.714271 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:18:04 crc kubenswrapper[4715]: E1203 22:18:04.714831 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6143cec8-8a39-443c-b850-109e24ac1200" containerName="glance-httpd" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.714850 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6143cec8-8a39-443c-b850-109e24ac1200" containerName="glance-httpd" Dec 03 22:18:04 crc kubenswrapper[4715]: E1203 22:18:04.714869 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6143cec8-8a39-443c-b850-109e24ac1200" containerName="glance-log" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.714876 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6143cec8-8a39-443c-b850-109e24ac1200" containerName="glance-log" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.715113 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6143cec8-8a39-443c-b850-109e24ac1200" containerName="glance-log" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.715143 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6143cec8-8a39-443c-b850-109e24ac1200" containerName="glance-httpd" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.716672 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.718611 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.718761 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.733438 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.841977 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.842231 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.842252 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-config-data\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.842304 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99136b0c-47a6-4503-806e-ec70fa8c2409-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.842331 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-scripts\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.842355 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.842436 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99136b0c-47a6-4503-806e-ec70fa8c2409-logs\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.842453 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnll8\" (UniqueName: \"kubernetes.io/projected/99136b0c-47a6-4503-806e-ec70fa8c2409-kube-api-access-xnll8\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.944842 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99136b0c-47a6-4503-806e-ec70fa8c2409-logs\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.944882 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnll8\" (UniqueName: \"kubernetes.io/projected/99136b0c-47a6-4503-806e-ec70fa8c2409-kube-api-access-xnll8\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.944957 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.944977 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.944993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-config-data\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.945029 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99136b0c-47a6-4503-806e-ec70fa8c2409-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.945427 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99136b0c-47a6-4503-806e-ec70fa8c2409-logs\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.945579 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99136b0c-47a6-4503-806e-ec70fa8c2409-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.945594 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-scripts\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.945694 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.946039 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.950659 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.950914 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.950949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-config-data\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.951039 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99136b0c-47a6-4503-806e-ec70fa8c2409-scripts\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.962997 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnll8\" (UniqueName: \"kubernetes.io/projected/99136b0c-47a6-4503-806e-ec70fa8c2409-kube-api-access-xnll8\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:04 crc kubenswrapper[4715]: I1203 22:18:04.971775 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"99136b0c-47a6-4503-806e-ec70fa8c2409\") " pod="openstack/glance-default-external-api-0" Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.034374 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.159542 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.159597 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.460011 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdwx4" event={"ID":"9fe015d5-6916-4c9a-a20f-d8be799a0e4c","Type":"ContainerStarted","Data":"9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809"} Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.463873 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerStarted","Data":"2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98"} Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.464495 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.467061 4715 generic.go:334] "Generic (PLEG): container finished" podID="279fe539-9749-4528-b219-df6323816ea7" containerID="f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459" exitCode=143 Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.467127 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"279fe539-9749-4528-b219-df6323816ea7","Type":"ContainerDied","Data":"f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459"} Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.481344 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pdwx4" podStartSLOduration=2.427386542 podStartE2EDuration="9.481325923s" podCreationTimestamp="2025-12-03 22:17:56 +0000 UTC" firstStartedPulling="2025-12-03 22:17:57.360926133 +0000 UTC m=+2174.103636738" lastFinishedPulling="2025-12-03 22:18:04.414865524 +0000 UTC m=+2181.157576119" observedRunningTime="2025-12-03 22:18:05.47932572 +0000 UTC m=+2182.222036315" watchObservedRunningTime="2025-12-03 22:18:05.481325923 +0000 UTC m=+2182.224036528" Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.505432 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.571007137 podStartE2EDuration="8.505417238s" podCreationTimestamp="2025-12-03 22:17:57 +0000 UTC" firstStartedPulling="2025-12-03 22:17:58.520493973 +0000 UTC m=+2175.263204568" lastFinishedPulling="2025-12-03 22:18:04.454904074 +0000 UTC m=+2181.197614669" observedRunningTime="2025-12-03 22:18:05.49836699 +0000 UTC m=+2182.241077605" watchObservedRunningTime="2025-12-03 22:18:05.505417238 +0000 UTC m=+2182.248127833" Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.646144 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6143cec8-8a39-443c-b850-109e24ac1200" path="/var/lib/kubelet/pods/6143cec8-8a39-443c-b850-109e24ac1200/volumes" Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.646948 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 22:18:05 crc kubenswrapper[4715]: I1203 22:18:05.851018 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:18:06 crc kubenswrapper[4715]: I1203 22:18:06.482970 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"99136b0c-47a6-4503-806e-ec70fa8c2409","Type":"ContainerStarted","Data":"5e6544565154f99d1092642ade5300e177a1f9584ebf71acc969b7257edcf5a1"} Dec 03 22:18:06 crc kubenswrapper[4715]: I1203 22:18:06.483267 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"99136b0c-47a6-4503-806e-ec70fa8c2409","Type":"ContainerStarted","Data":"0db4d87765a5933d9ac703cc1a9b9e417a591aaa1990c5768b4ec7a715ac6633"} Dec 03 22:18:06 crc kubenswrapper[4715]: I1203 22:18:06.565954 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:18:06 crc kubenswrapper[4715]: I1203 22:18:06.566035 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:18:06 crc kubenswrapper[4715]: I1203 22:18:06.630560 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.493534 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"99136b0c-47a6-4503-806e-ec70fa8c2409","Type":"ContainerStarted","Data":"7964323fc553f89fccbed1f1f16ff2f86802094e828165628c450a67fedafad5"} Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.493871 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="ceilometer-notification-agent" containerID="cri-o://9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8" gracePeriod=30 Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.493796 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="sg-core" containerID="cri-o://a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0" gracePeriod=30 Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.493781 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="proxy-httpd" containerID="cri-o://2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98" gracePeriod=30 Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.493738 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="ceilometer-central-agent" containerID="cri-o://539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d" gracePeriod=30 Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.496424 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-9nwdg"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.500944 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.513133 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-9nwdg"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.558715 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.558696598 podStartE2EDuration="3.558696598s" podCreationTimestamp="2025-12-03 22:18:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:18:07.554712171 +0000 UTC m=+2184.297422766" watchObservedRunningTime="2025-12-03 22:18:07.558696598 +0000 UTC m=+2184.301407193" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.626252 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-6ql9s"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.628077 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.659260 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-181b-account-create-update-c86tf"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.663012 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.669321 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.699262 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6ql9s"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.700069 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba52bb7b-0acf-4954-b20c-34b35f5b0169-operator-scripts\") pod \"nova-api-db-create-9nwdg\" (UID: \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\") " pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.700178 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvn9g\" (UniqueName: \"kubernetes.io/projected/ba52bb7b-0acf-4954-b20c-34b35f5b0169-kube-api-access-vvn9g\") pod \"nova-api-db-create-9nwdg\" (UID: \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\") " pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.709447 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-181b-account-create-update-c86tf"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.754215 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-4nhvz"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.755948 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.764672 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4nhvz"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.801872 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946473d7-6235-4989-b8b5-8bf32fac8873-operator-scripts\") pod \"nova-api-181b-account-create-update-c86tf\" (UID: \"946473d7-6235-4989-b8b5-8bf32fac8873\") " pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.801922 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddcgz\" (UniqueName: \"kubernetes.io/projected/946473d7-6235-4989-b8b5-8bf32fac8873-kube-api-access-ddcgz\") pod \"nova-api-181b-account-create-update-c86tf\" (UID: \"946473d7-6235-4989-b8b5-8bf32fac8873\") " pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.801947 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvn9g\" (UniqueName: \"kubernetes.io/projected/ba52bb7b-0acf-4954-b20c-34b35f5b0169-kube-api-access-vvn9g\") pod \"nova-api-db-create-9nwdg\" (UID: \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\") " pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.802039 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba52bb7b-0acf-4954-b20c-34b35f5b0169-operator-scripts\") pod \"nova-api-db-create-9nwdg\" (UID: \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\") " pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.802067 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-operator-scripts\") pod \"nova-cell0-db-create-6ql9s\" (UID: \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\") " pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.802126 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9swp\" (UniqueName: \"kubernetes.io/projected/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-kube-api-access-r9swp\") pod \"nova-cell0-db-create-6ql9s\" (UID: \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\") " pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.804825 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba52bb7b-0acf-4954-b20c-34b35f5b0169-operator-scripts\") pod \"nova-api-db-create-9nwdg\" (UID: \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\") " pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.820169 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-d43a-account-create-update-dscvc"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.821509 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.822938 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.830937 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvn9g\" (UniqueName: \"kubernetes.io/projected/ba52bb7b-0acf-4954-b20c-34b35f5b0169-kube-api-access-vvn9g\") pod \"nova-api-db-create-9nwdg\" (UID: \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\") " pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.837171 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d43a-account-create-update-dscvc"] Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.905846 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-operator-scripts\") pod \"nova-cell0-d43a-account-create-update-dscvc\" (UID: \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\") " pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.905901 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-operator-scripts\") pod \"nova-cell0-db-create-6ql9s\" (UID: \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\") " pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.905949 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9swp\" (UniqueName: \"kubernetes.io/projected/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-kube-api-access-r9swp\") pod \"nova-cell0-db-create-6ql9s\" (UID: \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\") " pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.905973 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kw98\" (UniqueName: \"kubernetes.io/projected/094b2293-50c2-4755-8abe-798dc10adb24-kube-api-access-4kw98\") pod \"nova-cell1-db-create-4nhvz\" (UID: \"094b2293-50c2-4755-8abe-798dc10adb24\") " pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.906015 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8c2r\" (UniqueName: \"kubernetes.io/projected/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-kube-api-access-p8c2r\") pod \"nova-cell0-d43a-account-create-update-dscvc\" (UID: \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\") " pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.906052 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/094b2293-50c2-4755-8abe-798dc10adb24-operator-scripts\") pod \"nova-cell1-db-create-4nhvz\" (UID: \"094b2293-50c2-4755-8abe-798dc10adb24\") " pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.906093 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946473d7-6235-4989-b8b5-8bf32fac8873-operator-scripts\") pod \"nova-api-181b-account-create-update-c86tf\" (UID: \"946473d7-6235-4989-b8b5-8bf32fac8873\") " pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.906119 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddcgz\" (UniqueName: \"kubernetes.io/projected/946473d7-6235-4989-b8b5-8bf32fac8873-kube-api-access-ddcgz\") pod \"nova-api-181b-account-create-update-c86tf\" (UID: \"946473d7-6235-4989-b8b5-8bf32fac8873\") " pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.911545 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946473d7-6235-4989-b8b5-8bf32fac8873-operator-scripts\") pod \"nova-api-181b-account-create-update-c86tf\" (UID: \"946473d7-6235-4989-b8b5-8bf32fac8873\") " pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.913354 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-operator-scripts\") pod \"nova-cell0-db-create-6ql9s\" (UID: \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\") " pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.926879 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9swp\" (UniqueName: \"kubernetes.io/projected/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-kube-api-access-r9swp\") pod \"nova-cell0-db-create-6ql9s\" (UID: \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\") " pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:07 crc kubenswrapper[4715]: I1203 22:18:07.935354 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddcgz\" (UniqueName: \"kubernetes.io/projected/946473d7-6235-4989-b8b5-8bf32fac8873-kube-api-access-ddcgz\") pod \"nova-api-181b-account-create-update-c86tf\" (UID: \"946473d7-6235-4989-b8b5-8bf32fac8873\") " pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:07.999668 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.007609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-operator-scripts\") pod \"nova-cell0-d43a-account-create-update-dscvc\" (UID: \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\") " pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.007690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kw98\" (UniqueName: \"kubernetes.io/projected/094b2293-50c2-4755-8abe-798dc10adb24-kube-api-access-4kw98\") pod \"nova-cell1-db-create-4nhvz\" (UID: \"094b2293-50c2-4755-8abe-798dc10adb24\") " pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.007734 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8c2r\" (UniqueName: \"kubernetes.io/projected/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-kube-api-access-p8c2r\") pod \"nova-cell0-d43a-account-create-update-dscvc\" (UID: \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\") " pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.007791 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/094b2293-50c2-4755-8abe-798dc10adb24-operator-scripts\") pod \"nova-cell1-db-create-4nhvz\" (UID: \"094b2293-50c2-4755-8abe-798dc10adb24\") " pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.008985 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-operator-scripts\") pod \"nova-cell0-d43a-account-create-update-dscvc\" (UID: \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\") " pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.010042 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/094b2293-50c2-4755-8abe-798dc10adb24-operator-scripts\") pod \"nova-cell1-db-create-4nhvz\" (UID: \"094b2293-50c2-4755-8abe-798dc10adb24\") " pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.013233 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-46c1-account-create-update-vfm2h"] Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.016294 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.023704 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.038902 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-46c1-account-create-update-vfm2h"] Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.044893 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kw98\" (UniqueName: \"kubernetes.io/projected/094b2293-50c2-4755-8abe-798dc10adb24-kube-api-access-4kw98\") pod \"nova-cell1-db-create-4nhvz\" (UID: \"094b2293-50c2-4755-8abe-798dc10adb24\") " pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.063720 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8c2r\" (UniqueName: \"kubernetes.io/projected/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-kube-api-access-p8c2r\") pod \"nova-cell0-d43a-account-create-update-dscvc\" (UID: \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\") " pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.114925 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9bd7ccb-3d1b-4621-89e8-96751e61be26-operator-scripts\") pod \"nova-cell1-46c1-account-create-update-vfm2h\" (UID: \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\") " pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.115489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlw22\" (UniqueName: \"kubernetes.io/projected/e9bd7ccb-3d1b-4621-89e8-96751e61be26-kube-api-access-qlw22\") pod \"nova-cell1-46c1-account-create-update-vfm2h\" (UID: \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\") " pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.117578 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.138928 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.189906 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.199291 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.220019 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlw22\" (UniqueName: \"kubernetes.io/projected/e9bd7ccb-3d1b-4621-89e8-96751e61be26-kube-api-access-qlw22\") pod \"nova-cell1-46c1-account-create-update-vfm2h\" (UID: \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\") " pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.220076 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9bd7ccb-3d1b-4621-89e8-96751e61be26-operator-scripts\") pod \"nova-cell1-46c1-account-create-update-vfm2h\" (UID: \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\") " pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.220778 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9bd7ccb-3d1b-4621-89e8-96751e61be26-operator-scripts\") pod \"nova-cell1-46c1-account-create-update-vfm2h\" (UID: \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\") " pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.245370 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlw22\" (UniqueName: \"kubernetes.io/projected/e9bd7ccb-3d1b-4621-89e8-96751e61be26-kube-api-access-qlw22\") pod \"nova-cell1-46c1-account-create-update-vfm2h\" (UID: \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\") " pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.260792 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.356544 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.424515 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8lw5\" (UniqueName: \"kubernetes.io/projected/279fe539-9749-4528-b219-df6323816ea7-kube-api-access-p8lw5\") pod \"279fe539-9749-4528-b219-df6323816ea7\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.424558 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-internal-tls-certs\") pod \"279fe539-9749-4528-b219-df6323816ea7\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.425068 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-httpd-run\") pod \"279fe539-9749-4528-b219-df6323816ea7\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.425092 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-logs\") pod \"279fe539-9749-4528-b219-df6323816ea7\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.425224 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"279fe539-9749-4528-b219-df6323816ea7\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.425305 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-combined-ca-bundle\") pod \"279fe539-9749-4528-b219-df6323816ea7\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.425375 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-scripts\") pod \"279fe539-9749-4528-b219-df6323816ea7\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.425417 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-config-data\") pod \"279fe539-9749-4528-b219-df6323816ea7\" (UID: \"279fe539-9749-4528-b219-df6323816ea7\") " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.425759 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "279fe539-9749-4528-b219-df6323816ea7" (UID: "279fe539-9749-4528-b219-df6323816ea7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.426063 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-logs" (OuterVolumeSpecName: "logs") pod "279fe539-9749-4528-b219-df6323816ea7" (UID: "279fe539-9749-4528-b219-df6323816ea7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.429490 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/279fe539-9749-4528-b219-df6323816ea7-kube-api-access-p8lw5" (OuterVolumeSpecName: "kube-api-access-p8lw5") pod "279fe539-9749-4528-b219-df6323816ea7" (UID: "279fe539-9749-4528-b219-df6323816ea7"). InnerVolumeSpecName "kube-api-access-p8lw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.429617 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "279fe539-9749-4528-b219-df6323816ea7" (UID: "279fe539-9749-4528-b219-df6323816ea7"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.435698 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-scripts" (OuterVolumeSpecName: "scripts") pod "279fe539-9749-4528-b219-df6323816ea7" (UID: "279fe539-9749-4528-b219-df6323816ea7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.458972 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "279fe539-9749-4528-b219-df6323816ea7" (UID: "279fe539-9749-4528-b219-df6323816ea7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.489784 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-config-data" (OuterVolumeSpecName: "config-data") pod "279fe539-9749-4528-b219-df6323816ea7" (UID: "279fe539-9749-4528-b219-df6323816ea7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.506416 4715 generic.go:334] "Generic (PLEG): container finished" podID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerID="2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98" exitCode=0 Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.506447 4715 generic.go:334] "Generic (PLEG): container finished" podID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerID="a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0" exitCode=2 Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.506479 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerDied","Data":"2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98"} Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.506516 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerDied","Data":"a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0"} Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.508373 4715 generic.go:334] "Generic (PLEG): container finished" podID="279fe539-9749-4528-b219-df6323816ea7" containerID="a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f" exitCode=0 Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.509227 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.509928 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"279fe539-9749-4528-b219-df6323816ea7","Type":"ContainerDied","Data":"a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f"} Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.509958 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"279fe539-9749-4528-b219-df6323816ea7","Type":"ContainerDied","Data":"176af77e30c5143c1fc7b70bcecfab63b2a76e856708458ea11029e0d24f963e"} Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.510002 4715 scope.go:117] "RemoveContainer" containerID="a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.527932 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "279fe539-9749-4528-b219-df6323816ea7" (UID: "279fe539-9749-4528-b219-df6323816ea7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.528195 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.528274 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.528328 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8lw5\" (UniqueName: \"kubernetes.io/projected/279fe539-9749-4528-b219-df6323816ea7-kube-api-access-p8lw5\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.528384 4715 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.528437 4715 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.528494 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/279fe539-9749-4528-b219-df6323816ea7-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.528655 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.528763 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/279fe539-9749-4528-b219-df6323816ea7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.536686 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-6ql9s"] Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.547470 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.577934 4715 scope.go:117] "RemoveContainer" containerID="f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459" Dec 03 22:18:08 crc kubenswrapper[4715]: W1203 22:18:08.585161 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a3f66d8_7bbb_4939_bef3_ebb7365e53a2.slice/crio-eae918b67cb4fde26618e0af39348c32fbc50aba0c11ae65a3dacc52afab0bd7 WatchSource:0}: Error finding container eae918b67cb4fde26618e0af39348c32fbc50aba0c11ae65a3dacc52afab0bd7: Status 404 returned error can't find the container with id eae918b67cb4fde26618e0af39348c32fbc50aba0c11ae65a3dacc52afab0bd7 Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.598980 4715 scope.go:117] "RemoveContainer" containerID="a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f" Dec 03 22:18:08 crc kubenswrapper[4715]: E1203 22:18:08.599448 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f\": container with ID starting with a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f not found: ID does not exist" containerID="a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.599517 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f"} err="failed to get container status \"a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f\": rpc error: code = NotFound desc = could not find container \"a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f\": container with ID starting with a7de422504eeb7f8edfcdc58067a2a5a5f4ed79e5d1f08a7b41ee0551c0e418f not found: ID does not exist" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.599552 4715 scope.go:117] "RemoveContainer" containerID="f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459" Dec 03 22:18:08 crc kubenswrapper[4715]: E1203 22:18:08.599943 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459\": container with ID starting with f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459 not found: ID does not exist" containerID="f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.600022 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459"} err="failed to get container status \"f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459\": rpc error: code = NotFound desc = could not find container \"f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459\": container with ID starting with f21517c1fad5206298429f598088ed2fe5cefcbd34156d29c446af82bd309459 not found: ID does not exist" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.630923 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.789669 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-181b-account-create-update-c86tf"] Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.800851 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4nhvz"] Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.821465 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-9nwdg"] Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.915173 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d43a-account-create-update-dscvc"] Dec 03 22:18:08 crc kubenswrapper[4715]: W1203 22:18:08.920075 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc39bbab4_12df_47cb_b3f2_85e2d2f8cac9.slice/crio-8ac88903f04e0498d5279f8624412d9703008db3c5fec66d67c85530278c79e4 WatchSource:0}: Error finding container 8ac88903f04e0498d5279f8624412d9703008db3c5fec66d67c85530278c79e4: Status 404 returned error can't find the container with id 8ac88903f04e0498d5279f8624412d9703008db3c5fec66d67c85530278c79e4 Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.970528 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:18:08 crc kubenswrapper[4715]: I1203 22:18:08.985762 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.001958 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:18:09 crc kubenswrapper[4715]: E1203 22:18:09.003060 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279fe539-9749-4528-b219-df6323816ea7" containerName="glance-log" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.003081 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="279fe539-9749-4528-b219-df6323816ea7" containerName="glance-log" Dec 03 22:18:09 crc kubenswrapper[4715]: E1203 22:18:09.003124 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="279fe539-9749-4528-b219-df6323816ea7" containerName="glance-httpd" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.003132 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="279fe539-9749-4528-b219-df6323816ea7" containerName="glance-httpd" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.003636 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="279fe539-9749-4528-b219-df6323816ea7" containerName="glance-log" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.003672 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="279fe539-9749-4528-b219-df6323816ea7" containerName="glance-httpd" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.005436 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.010989 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.012280 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.019044 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.072734 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-46c1-account-create-update-vfm2h"] Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.157221 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.157328 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.157349 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/20ba12ff-26f9-4100-b21d-0545d33fc7d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.157378 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.157585 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20ba12ff-26f9-4100-b21d-0545d33fc7d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.157690 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.157897 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lw6b\" (UniqueName: \"kubernetes.io/projected/20ba12ff-26f9-4100-b21d-0545d33fc7d1-kube-api-access-8lw6b\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.157936 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259565 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lw6b\" (UniqueName: \"kubernetes.io/projected/20ba12ff-26f9-4100-b21d-0545d33fc7d1-kube-api-access-8lw6b\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259618 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259658 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259710 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259728 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/20ba12ff-26f9-4100-b21d-0545d33fc7d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259753 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259788 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20ba12ff-26f9-4100-b21d-0545d33fc7d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259813 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.259920 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.261407 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20ba12ff-26f9-4100-b21d-0545d33fc7d1-logs\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.261456 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/20ba12ff-26f9-4100-b21d-0545d33fc7d1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.268705 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.270549 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.271682 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.271750 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/20ba12ff-26f9-4100-b21d-0545d33fc7d1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.278431 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lw6b\" (UniqueName: \"kubernetes.io/projected/20ba12ff-26f9-4100-b21d-0545d33fc7d1-kube-api-access-8lw6b\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.297761 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"20ba12ff-26f9-4100-b21d-0545d33fc7d1\") " pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.343884 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.522206 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" event={"ID":"e9bd7ccb-3d1b-4621-89e8-96751e61be26","Type":"ContainerStarted","Data":"9a977b94eb96818749cc9e015af6cc01f2febf91938634db09e45a5eb5715ecd"} Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.525940 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-181b-account-create-update-c86tf" event={"ID":"946473d7-6235-4989-b8b5-8bf32fac8873","Type":"ContainerStarted","Data":"ef3ca5254e7c605bb53733cf904e1dd46b522b1afbccd80f55c1ef8fe9b3823b"} Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.530747 4715 generic.go:334] "Generic (PLEG): container finished" podID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerID="9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8" exitCode=0 Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.530803 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerDied","Data":"9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8"} Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.532344 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6ql9s" event={"ID":"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2","Type":"ContainerStarted","Data":"eae918b67cb4fde26618e0af39348c32fbc50aba0c11ae65a3dacc52afab0bd7"} Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.535289 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d43a-account-create-update-dscvc" event={"ID":"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9","Type":"ContainerStarted","Data":"8ac88903f04e0498d5279f8624412d9703008db3c5fec66d67c85530278c79e4"} Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.537622 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4nhvz" event={"ID":"094b2293-50c2-4755-8abe-798dc10adb24","Type":"ContainerStarted","Data":"f1db0c342f0fec4a7eade13e6233b1efd73944b6af4ca3bd681306413c4c8dc4"} Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.538471 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9nwdg" event={"ID":"ba52bb7b-0acf-4954-b20c-34b35f5b0169","Type":"ContainerStarted","Data":"49cacd5bc59d61221e61d06c7750057b4659e2cce19b704ced9cc1d351ce531c"} Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.652727 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="279fe539-9749-4528-b219-df6323816ea7" path="/var/lib/kubelet/pods/279fe539-9749-4528-b219-df6323816ea7/volumes" Dec 03 22:18:09 crc kubenswrapper[4715]: I1203 22:18:09.928763 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.547652 4715 generic.go:334] "Generic (PLEG): container finished" podID="094b2293-50c2-4755-8abe-798dc10adb24" containerID="a4406c2da4437418f16f953d3850ca7a0bc3619e2d1189b88ed76dbe28c67e27" exitCode=0 Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.548808 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4nhvz" event={"ID":"094b2293-50c2-4755-8abe-798dc10adb24","Type":"ContainerDied","Data":"a4406c2da4437418f16f953d3850ca7a0bc3619e2d1189b88ed76dbe28c67e27"} Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.553026 4715 generic.go:334] "Generic (PLEG): container finished" podID="ba52bb7b-0acf-4954-b20c-34b35f5b0169" containerID="fb011ef848a00c81e2da1976439fb931a33b9cae1948a954793a73aa15301343" exitCode=0 Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.553081 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9nwdg" event={"ID":"ba52bb7b-0acf-4954-b20c-34b35f5b0169","Type":"ContainerDied","Data":"fb011ef848a00c81e2da1976439fb931a33b9cae1948a954793a73aa15301343"} Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.554248 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"20ba12ff-26f9-4100-b21d-0545d33fc7d1","Type":"ContainerStarted","Data":"f1e2252c47c52f21ae4cb9fcf27f22677c230245d5fe526777a9b8e6df2697c1"} Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.556663 4715 generic.go:334] "Generic (PLEG): container finished" podID="e9bd7ccb-3d1b-4621-89e8-96751e61be26" containerID="84743759692db45134fdbe7f3474a79c4ce13731927ce11d290a5e950822455e" exitCode=0 Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.556707 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" event={"ID":"e9bd7ccb-3d1b-4621-89e8-96751e61be26","Type":"ContainerDied","Data":"84743759692db45134fdbe7f3474a79c4ce13731927ce11d290a5e950822455e"} Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.558225 4715 generic.go:334] "Generic (PLEG): container finished" podID="946473d7-6235-4989-b8b5-8bf32fac8873" containerID="f606fae52baae301c32649dab26448703d133e0c6c55cf5480176ebea303d755" exitCode=0 Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.558275 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-181b-account-create-update-c86tf" event={"ID":"946473d7-6235-4989-b8b5-8bf32fac8873","Type":"ContainerDied","Data":"f606fae52baae301c32649dab26448703d133e0c6c55cf5480176ebea303d755"} Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.559939 4715 generic.go:334] "Generic (PLEG): container finished" podID="9a3f66d8-7bbb-4939-bef3-ebb7365e53a2" containerID="9b14ea35554400954df1f9303ee8433d233d9452cb383b06cf583b15be7d7d8d" exitCode=0 Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.559982 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6ql9s" event={"ID":"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2","Type":"ContainerDied","Data":"9b14ea35554400954df1f9303ee8433d233d9452cb383b06cf583b15be7d7d8d"} Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.561347 4715 generic.go:334] "Generic (PLEG): container finished" podID="c39bbab4-12df-47cb-b3f2-85e2d2f8cac9" containerID="42708c24c3bcec216fe4dd11b4853a7f1bd27b13321c4b02aa48f63a9130ef0d" exitCode=0 Dec 03 22:18:10 crc kubenswrapper[4715]: I1203 22:18:10.561373 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d43a-account-create-update-dscvc" event={"ID":"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9","Type":"ContainerDied","Data":"42708c24c3bcec216fe4dd11b4853a7f1bd27b13321c4b02aa48f63a9130ef0d"} Dec 03 22:18:11 crc kubenswrapper[4715]: I1203 22:18:11.575840 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"20ba12ff-26f9-4100-b21d-0545d33fc7d1","Type":"ContainerStarted","Data":"b0b0b7edc24fa7f52de34c42232b65277c4b0f7e9489fef8ec416262b46b553e"} Dec 03 22:18:11 crc kubenswrapper[4715]: I1203 22:18:11.576155 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"20ba12ff-26f9-4100-b21d-0545d33fc7d1","Type":"ContainerStarted","Data":"6b8f25ea0a3e0ab4b804ceb793dede1ac029afc3413412e5ba92a598e46a367c"} Dec 03 22:18:11 crc kubenswrapper[4715]: I1203 22:18:11.606293 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.606269481 podStartE2EDuration="3.606269481s" podCreationTimestamp="2025-12-03 22:18:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:18:11.60252853 +0000 UTC m=+2188.345239135" watchObservedRunningTime="2025-12-03 22:18:11.606269481 +0000 UTC m=+2188.348980076" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.008367 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.142863 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9bd7ccb-3d1b-4621-89e8-96751e61be26-operator-scripts\") pod \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\" (UID: \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.142934 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlw22\" (UniqueName: \"kubernetes.io/projected/e9bd7ccb-3d1b-4621-89e8-96751e61be26-kube-api-access-qlw22\") pod \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\" (UID: \"e9bd7ccb-3d1b-4621-89e8-96751e61be26\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.143684 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9bd7ccb-3d1b-4621-89e8-96751e61be26-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e9bd7ccb-3d1b-4621-89e8-96751e61be26" (UID: "e9bd7ccb-3d1b-4621-89e8-96751e61be26"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.150315 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9bd7ccb-3d1b-4621-89e8-96751e61be26-kube-api-access-qlw22" (OuterVolumeSpecName: "kube-api-access-qlw22") pod "e9bd7ccb-3d1b-4621-89e8-96751e61be26" (UID: "e9bd7ccb-3d1b-4621-89e8-96751e61be26"). InnerVolumeSpecName "kube-api-access-qlw22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.232477 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.242555 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.250205 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.250208 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e9bd7ccb-3d1b-4621-89e8-96751e61be26-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.251725 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlw22\" (UniqueName: \"kubernetes.io/projected/e9bd7ccb-3d1b-4621-89e8-96751e61be26-kube-api-access-qlw22\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.264220 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.285615 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.352782 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddcgz\" (UniqueName: \"kubernetes.io/projected/946473d7-6235-4989-b8b5-8bf32fac8873-kube-api-access-ddcgz\") pod \"946473d7-6235-4989-b8b5-8bf32fac8873\" (UID: \"946473d7-6235-4989-b8b5-8bf32fac8873\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.352899 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-operator-scripts\") pod \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\" (UID: \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.352936 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8c2r\" (UniqueName: \"kubernetes.io/projected/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-kube-api-access-p8c2r\") pod \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\" (UID: \"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.352992 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946473d7-6235-4989-b8b5-8bf32fac8873-operator-scripts\") pod \"946473d7-6235-4989-b8b5-8bf32fac8873\" (UID: \"946473d7-6235-4989-b8b5-8bf32fac8873\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.353064 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kw98\" (UniqueName: \"kubernetes.io/projected/094b2293-50c2-4755-8abe-798dc10adb24-kube-api-access-4kw98\") pod \"094b2293-50c2-4755-8abe-798dc10adb24\" (UID: \"094b2293-50c2-4755-8abe-798dc10adb24\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.353198 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/094b2293-50c2-4755-8abe-798dc10adb24-operator-scripts\") pod \"094b2293-50c2-4755-8abe-798dc10adb24\" (UID: \"094b2293-50c2-4755-8abe-798dc10adb24\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.354114 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/094b2293-50c2-4755-8abe-798dc10adb24-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "094b2293-50c2-4755-8abe-798dc10adb24" (UID: "094b2293-50c2-4755-8abe-798dc10adb24"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.354128 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946473d7-6235-4989-b8b5-8bf32fac8873-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "946473d7-6235-4989-b8b5-8bf32fac8873" (UID: "946473d7-6235-4989-b8b5-8bf32fac8873"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.354715 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c39bbab4-12df-47cb-b3f2-85e2d2f8cac9" (UID: "c39bbab4-12df-47cb-b3f2-85e2d2f8cac9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.357746 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946473d7-6235-4989-b8b5-8bf32fac8873-kube-api-access-ddcgz" (OuterVolumeSpecName: "kube-api-access-ddcgz") pod "946473d7-6235-4989-b8b5-8bf32fac8873" (UID: "946473d7-6235-4989-b8b5-8bf32fac8873"). InnerVolumeSpecName "kube-api-access-ddcgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.357841 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/094b2293-50c2-4755-8abe-798dc10adb24-kube-api-access-4kw98" (OuterVolumeSpecName: "kube-api-access-4kw98") pod "094b2293-50c2-4755-8abe-798dc10adb24" (UID: "094b2293-50c2-4755-8abe-798dc10adb24"). InnerVolumeSpecName "kube-api-access-4kw98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.360614 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-kube-api-access-p8c2r" (OuterVolumeSpecName: "kube-api-access-p8c2r") pod "c39bbab4-12df-47cb-b3f2-85e2d2f8cac9" (UID: "c39bbab4-12df-47cb-b3f2-85e2d2f8cac9"). InnerVolumeSpecName "kube-api-access-p8c2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.454896 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvn9g\" (UniqueName: \"kubernetes.io/projected/ba52bb7b-0acf-4954-b20c-34b35f5b0169-kube-api-access-vvn9g\") pod \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\" (UID: \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.455195 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9swp\" (UniqueName: \"kubernetes.io/projected/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-kube-api-access-r9swp\") pod \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\" (UID: \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.455240 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba52bb7b-0acf-4954-b20c-34b35f5b0169-operator-scripts\") pod \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\" (UID: \"ba52bb7b-0acf-4954-b20c-34b35f5b0169\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.455333 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-operator-scripts\") pod \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\" (UID: \"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2\") " Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.455991 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.456017 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8c2r\" (UniqueName: \"kubernetes.io/projected/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9-kube-api-access-p8c2r\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.456033 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946473d7-6235-4989-b8b5-8bf32fac8873-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.456048 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kw98\" (UniqueName: \"kubernetes.io/projected/094b2293-50c2-4755-8abe-798dc10adb24-kube-api-access-4kw98\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.456059 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/094b2293-50c2-4755-8abe-798dc10adb24-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.456075 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddcgz\" (UniqueName: \"kubernetes.io/projected/946473d7-6235-4989-b8b5-8bf32fac8873-kube-api-access-ddcgz\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.456480 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9a3f66d8-7bbb-4939-bef3-ebb7365e53a2" (UID: "9a3f66d8-7bbb-4939-bef3-ebb7365e53a2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.457962 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba52bb7b-0acf-4954-b20c-34b35f5b0169-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ba52bb7b-0acf-4954-b20c-34b35f5b0169" (UID: "ba52bb7b-0acf-4954-b20c-34b35f5b0169"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.461923 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-kube-api-access-r9swp" (OuterVolumeSpecName: "kube-api-access-r9swp") pod "9a3f66d8-7bbb-4939-bef3-ebb7365e53a2" (UID: "9a3f66d8-7bbb-4939-bef3-ebb7365e53a2"). InnerVolumeSpecName "kube-api-access-r9swp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.462067 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba52bb7b-0acf-4954-b20c-34b35f5b0169-kube-api-access-vvn9g" (OuterVolumeSpecName: "kube-api-access-vvn9g") pod "ba52bb7b-0acf-4954-b20c-34b35f5b0169" (UID: "ba52bb7b-0acf-4954-b20c-34b35f5b0169"). InnerVolumeSpecName "kube-api-access-vvn9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.557345 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.557376 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvn9g\" (UniqueName: \"kubernetes.io/projected/ba52bb7b-0acf-4954-b20c-34b35f5b0169-kube-api-access-vvn9g\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.557389 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9swp\" (UniqueName: \"kubernetes.io/projected/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2-kube-api-access-r9swp\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.557397 4715 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ba52bb7b-0acf-4954-b20c-34b35f5b0169-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.585940 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-181b-account-create-update-c86tf" event={"ID":"946473d7-6235-4989-b8b5-8bf32fac8873","Type":"ContainerDied","Data":"ef3ca5254e7c605bb53733cf904e1dd46b522b1afbccd80f55c1ef8fe9b3823b"} Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.585975 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef3ca5254e7c605bb53733cf904e1dd46b522b1afbccd80f55c1ef8fe9b3823b" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.586033 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-181b-account-create-update-c86tf" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.598557 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-6ql9s" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.598552 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-6ql9s" event={"ID":"9a3f66d8-7bbb-4939-bef3-ebb7365e53a2","Type":"ContainerDied","Data":"eae918b67cb4fde26618e0af39348c32fbc50aba0c11ae65a3dacc52afab0bd7"} Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.598767 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eae918b67cb4fde26618e0af39348c32fbc50aba0c11ae65a3dacc52afab0bd7" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.600217 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d43a-account-create-update-dscvc" event={"ID":"c39bbab4-12df-47cb-b3f2-85e2d2f8cac9","Type":"ContainerDied","Data":"8ac88903f04e0498d5279f8624412d9703008db3c5fec66d67c85530278c79e4"} Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.600254 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ac88903f04e0498d5279f8624412d9703008db3c5fec66d67c85530278c79e4" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.600360 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d43a-account-create-update-dscvc" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.603943 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-9nwdg" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.603999 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-9nwdg" event={"ID":"ba52bb7b-0acf-4954-b20c-34b35f5b0169","Type":"ContainerDied","Data":"49cacd5bc59d61221e61d06c7750057b4659e2cce19b704ced9cc1d351ce531c"} Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.604047 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49cacd5bc59d61221e61d06c7750057b4659e2cce19b704ced9cc1d351ce531c" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.611020 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4nhvz" event={"ID":"094b2293-50c2-4755-8abe-798dc10adb24","Type":"ContainerDied","Data":"f1db0c342f0fec4a7eade13e6233b1efd73944b6af4ca3bd681306413c4c8dc4"} Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.611052 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4nhvz" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.611060 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1db0c342f0fec4a7eade13e6233b1efd73944b6af4ca3bd681306413c4c8dc4" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.613694 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" event={"ID":"e9bd7ccb-3d1b-4621-89e8-96751e61be26","Type":"ContainerDied","Data":"9a977b94eb96818749cc9e015af6cc01f2febf91938634db09e45a5eb5715ecd"} Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.613708 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-46c1-account-create-update-vfm2h" Dec 03 22:18:12 crc kubenswrapper[4715]: I1203 22:18:12.613728 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a977b94eb96818749cc9e015af6cc01f2febf91938634db09e45a5eb5715ecd" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.481272 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.491659 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-sg-core-conf-yaml\") pod \"c851dcc7-cdd0-438f-ade3-a8d07274897b\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.491890 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-scripts\") pod \"c851dcc7-cdd0-438f-ade3-a8d07274897b\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.491985 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-combined-ca-bundle\") pod \"c851dcc7-cdd0-438f-ade3-a8d07274897b\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.492196 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xndf\" (UniqueName: \"kubernetes.io/projected/c851dcc7-cdd0-438f-ade3-a8d07274897b-kube-api-access-4xndf\") pod \"c851dcc7-cdd0-438f-ade3-a8d07274897b\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.492285 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-config-data\") pod \"c851dcc7-cdd0-438f-ade3-a8d07274897b\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.492446 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-run-httpd\") pod \"c851dcc7-cdd0-438f-ade3-a8d07274897b\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.492588 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-log-httpd\") pod \"c851dcc7-cdd0-438f-ade3-a8d07274897b\" (UID: \"c851dcc7-cdd0-438f-ade3-a8d07274897b\") " Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.493214 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c851dcc7-cdd0-438f-ade3-a8d07274897b" (UID: "c851dcc7-cdd0-438f-ade3-a8d07274897b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.493583 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c851dcc7-cdd0-438f-ade3-a8d07274897b" (UID: "c851dcc7-cdd0-438f-ade3-a8d07274897b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.497021 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-scripts" (OuterVolumeSpecName: "scripts") pod "c851dcc7-cdd0-438f-ade3-a8d07274897b" (UID: "c851dcc7-cdd0-438f-ade3-a8d07274897b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.498995 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c851dcc7-cdd0-438f-ade3-a8d07274897b-kube-api-access-4xndf" (OuterVolumeSpecName: "kube-api-access-4xndf") pod "c851dcc7-cdd0-438f-ade3-a8d07274897b" (UID: "c851dcc7-cdd0-438f-ade3-a8d07274897b"). InnerVolumeSpecName "kube-api-access-4xndf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.538616 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c851dcc7-cdd0-438f-ade3-a8d07274897b" (UID: "c851dcc7-cdd0-438f-ade3-a8d07274897b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.596466 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.596545 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c851dcc7-cdd0-438f-ade3-a8d07274897b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.596584 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.596600 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.596611 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xndf\" (UniqueName: \"kubernetes.io/projected/c851dcc7-cdd0-438f-ade3-a8d07274897b-kube-api-access-4xndf\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.609640 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c851dcc7-cdd0-438f-ade3-a8d07274897b" (UID: "c851dcc7-cdd0-438f-ade3-a8d07274897b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.635315 4715 generic.go:334] "Generic (PLEG): container finished" podID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerID="539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d" exitCode=0 Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.635370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerDied","Data":"539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d"} Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.635399 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c851dcc7-cdd0-438f-ade3-a8d07274897b","Type":"ContainerDied","Data":"dc2ff9d8c95013c7f5cab1c29b0c739c7670f37fc80835fa8a25df6edc18309c"} Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.635421 4715 scope.go:117] "RemoveContainer" containerID="2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.635581 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.654656 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-config-data" (OuterVolumeSpecName: "config-data") pod "c851dcc7-cdd0-438f-ade3-a8d07274897b" (UID: "c851dcc7-cdd0-438f-ade3-a8d07274897b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.656979 4715 scope.go:117] "RemoveContainer" containerID="a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.685873 4715 scope.go:117] "RemoveContainer" containerID="9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.698150 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.698177 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c851dcc7-cdd0-438f-ade3-a8d07274897b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.712640 4715 scope.go:117] "RemoveContainer" containerID="539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.738755 4715 scope.go:117] "RemoveContainer" containerID="2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.739079 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98\": container with ID starting with 2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98 not found: ID does not exist" containerID="2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.739112 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98"} err="failed to get container status \"2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98\": rpc error: code = NotFound desc = could not find container \"2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98\": container with ID starting with 2bae3b3a87a48d6a3fa8b5ac872c7ae542a5bfcce9a15278b75b5c21b5af8a98 not found: ID does not exist" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.739135 4715 scope.go:117] "RemoveContainer" containerID="a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.739342 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0\": container with ID starting with a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0 not found: ID does not exist" containerID="a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.739366 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0"} err="failed to get container status \"a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0\": rpc error: code = NotFound desc = could not find container \"a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0\": container with ID starting with a65a6652a49d1193fa95c5acce8b536d359261d4f1154760530ad0af6157f5a0 not found: ID does not exist" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.739378 4715 scope.go:117] "RemoveContainer" containerID="9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.739563 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8\": container with ID starting with 9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8 not found: ID does not exist" containerID="9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.739584 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8"} err="failed to get container status \"9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8\": rpc error: code = NotFound desc = could not find container \"9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8\": container with ID starting with 9358b7838a88b0c982127c8d104225d895933c4c96ab138e57e5e8b551c832b8 not found: ID does not exist" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.739595 4715 scope.go:117] "RemoveContainer" containerID="539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.739850 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d\": container with ID starting with 539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d not found: ID does not exist" containerID="539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.739870 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d"} err="failed to get container status \"539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d\": rpc error: code = NotFound desc = could not find container \"539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d\": container with ID starting with 539661d9e13869fce2a2903b7c36fd553667712270aa14429f912c1ac1f54e4d not found: ID does not exist" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.966621 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.974919 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.996205 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.996933 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="sg-core" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.997034 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="sg-core" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.997124 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="ceilometer-notification-agent" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.997201 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="ceilometer-notification-agent" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.997291 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946473d7-6235-4989-b8b5-8bf32fac8873" containerName="mariadb-account-create-update" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.997366 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="946473d7-6235-4989-b8b5-8bf32fac8873" containerName="mariadb-account-create-update" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.997442 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c39bbab4-12df-47cb-b3f2-85e2d2f8cac9" containerName="mariadb-account-create-update" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.997533 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c39bbab4-12df-47cb-b3f2-85e2d2f8cac9" containerName="mariadb-account-create-update" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.997610 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9bd7ccb-3d1b-4621-89e8-96751e61be26" containerName="mariadb-account-create-update" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.997688 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9bd7ccb-3d1b-4621-89e8-96751e61be26" containerName="mariadb-account-create-update" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.997770 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="proxy-httpd" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.997876 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="proxy-httpd" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.997971 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094b2293-50c2-4755-8abe-798dc10adb24" containerName="mariadb-database-create" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.998046 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="094b2293-50c2-4755-8abe-798dc10adb24" containerName="mariadb-database-create" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.998137 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="ceilometer-central-agent" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.998205 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="ceilometer-central-agent" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.998294 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3f66d8-7bbb-4939-bef3-ebb7365e53a2" containerName="mariadb-database-create" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.998386 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3f66d8-7bbb-4939-bef3-ebb7365e53a2" containerName="mariadb-database-create" Dec 03 22:18:14 crc kubenswrapper[4715]: E1203 22:18:14.998466 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba52bb7b-0acf-4954-b20c-34b35f5b0169" containerName="mariadb-database-create" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.999141 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba52bb7b-0acf-4954-b20c-34b35f5b0169" containerName="mariadb-database-create" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.999492 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="sg-core" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.999631 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="proxy-httpd" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.999718 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="946473d7-6235-4989-b8b5-8bf32fac8873" containerName="mariadb-account-create-update" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.999798 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba52bb7b-0acf-4954-b20c-34b35f5b0169" containerName="mariadb-database-create" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.999887 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9bd7ccb-3d1b-4621-89e8-96751e61be26" containerName="mariadb-account-create-update" Dec 03 22:18:14 crc kubenswrapper[4715]: I1203 22:18:14.999962 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="094b2293-50c2-4755-8abe-798dc10adb24" containerName="mariadb-database-create" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.000043 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3f66d8-7bbb-4939-bef3-ebb7365e53a2" containerName="mariadb-database-create" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.000111 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c39bbab4-12df-47cb-b3f2-85e2d2f8cac9" containerName="mariadb-account-create-update" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.000198 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="ceilometer-central-agent" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.000269 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" containerName="ceilometer-notification-agent" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.002303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.004420 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.004660 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.009679 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.035272 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.035315 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.064611 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.078159 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.108107 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.108188 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.108242 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnjfj\" (UniqueName: \"kubernetes.io/projected/d226be51-32f8-4710-af9f-afc1253b1485-kube-api-access-vnjfj\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.108311 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-scripts\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.108489 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-config-data\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.108616 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-run-httpd\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.108702 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-log-httpd\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.210034 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-log-httpd\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.210167 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.210204 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.210245 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnjfj\" (UniqueName: \"kubernetes.io/projected/d226be51-32f8-4710-af9f-afc1253b1485-kube-api-access-vnjfj\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.210288 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-scripts\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.210308 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-config-data\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.210347 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-run-httpd\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.210597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-log-httpd\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.213793 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-run-httpd\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.218423 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.220672 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-scripts\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.225210 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.226891 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-config-data\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.228927 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnjfj\" (UniqueName: \"kubernetes.io/projected/d226be51-32f8-4710-af9f-afc1253b1485-kube-api-access-vnjfj\") pod \"ceilometer-0\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.363625 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.648808 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c851dcc7-cdd0-438f-ade3-a8d07274897b" path="/var/lib/kubelet/pods/c851dcc7-cdd0-438f-ade3-a8d07274897b/volumes" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.649882 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.649907 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 22:18:15 crc kubenswrapper[4715]: W1203 22:18:15.875728 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd226be51_32f8_4710_af9f_afc1253b1485.slice/crio-2e923eadba35f98115a5b7508189cfc0a073013cec81feb42b049d7b904e8822 WatchSource:0}: Error finding container 2e923eadba35f98115a5b7508189cfc0a073013cec81feb42b049d7b904e8822: Status 404 returned error can't find the container with id 2e923eadba35f98115a5b7508189cfc0a073013cec81feb42b049d7b904e8822 Dec 03 22:18:15 crc kubenswrapper[4715]: I1203 22:18:15.879867 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:18:16 crc kubenswrapper[4715]: I1203 22:18:16.615900 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:18:16 crc kubenswrapper[4715]: I1203 22:18:16.658370 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerStarted","Data":"2298e6b06d0ef1f39ff8e2d48041a92c76c9b3b2c5632e91a003824acefbc015"} Dec 03 22:18:16 crc kubenswrapper[4715]: I1203 22:18:16.658442 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerStarted","Data":"2e923eadba35f98115a5b7508189cfc0a073013cec81feb42b049d7b904e8822"} Dec 03 22:18:16 crc kubenswrapper[4715]: I1203 22:18:16.674342 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pdwx4"] Dec 03 22:18:16 crc kubenswrapper[4715]: I1203 22:18:16.674587 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pdwx4" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerName="registry-server" containerID="cri-o://9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809" gracePeriod=2 Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.167934 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.352807 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-utilities\") pod \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.352855 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvdj6\" (UniqueName: \"kubernetes.io/projected/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-kube-api-access-zvdj6\") pod \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.352879 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-catalog-content\") pod \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\" (UID: \"9fe015d5-6916-4c9a-a20f-d8be799a0e4c\") " Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.353554 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-utilities" (OuterVolumeSpecName: "utilities") pod "9fe015d5-6916-4c9a-a20f-d8be799a0e4c" (UID: "9fe015d5-6916-4c9a-a20f-d8be799a0e4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.372680 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-kube-api-access-zvdj6" (OuterVolumeSpecName: "kube-api-access-zvdj6") pod "9fe015d5-6916-4c9a-a20f-d8be799a0e4c" (UID: "9fe015d5-6916-4c9a-a20f-d8be799a0e4c"). InnerVolumeSpecName "kube-api-access-zvdj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.401905 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9fe015d5-6916-4c9a-a20f-d8be799a0e4c" (UID: "9fe015d5-6916-4c9a-a20f-d8be799a0e4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.455421 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.455479 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvdj6\" (UniqueName: \"kubernetes.io/projected/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-kube-api-access-zvdj6\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.455495 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9fe015d5-6916-4c9a-a20f-d8be799a0e4c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.507144 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.524888 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.670369 4715 generic.go:334] "Generic (PLEG): container finished" podID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerID="9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809" exitCode=0 Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.671380 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdwx4" event={"ID":"9fe015d5-6916-4c9a-a20f-d8be799a0e4c","Type":"ContainerDied","Data":"9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809"} Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.671449 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pdwx4" event={"ID":"9fe015d5-6916-4c9a-a20f-d8be799a0e4c","Type":"ContainerDied","Data":"a167ef45883d0b839ac5a6f660c1162250a2383f6f9118e83ebbbf55c1acfe97"} Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.671471 4715 scope.go:117] "RemoveContainer" containerID="9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.672115 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pdwx4" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.705407 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pdwx4"] Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.707835 4715 scope.go:117] "RemoveContainer" containerID="0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.732045 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pdwx4"] Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.746964 4715 scope.go:117] "RemoveContainer" containerID="0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.796228 4715 scope.go:117] "RemoveContainer" containerID="9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809" Dec 03 22:18:17 crc kubenswrapper[4715]: E1203 22:18:17.796705 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809\": container with ID starting with 9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809 not found: ID does not exist" containerID="9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.796738 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809"} err="failed to get container status \"9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809\": rpc error: code = NotFound desc = could not find container \"9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809\": container with ID starting with 9b21d08b10a6390b1d279976b4ca51bdfab258eb62baf62b5ea812c021b2d809 not found: ID does not exist" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.796760 4715 scope.go:117] "RemoveContainer" containerID="0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6" Dec 03 22:18:17 crc kubenswrapper[4715]: E1203 22:18:17.797103 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6\": container with ID starting with 0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6 not found: ID does not exist" containerID="0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.797123 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6"} err="failed to get container status \"0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6\": rpc error: code = NotFound desc = could not find container \"0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6\": container with ID starting with 0a1d1b05c887f0c791a083241f219cc109416cef9db12effb65ced4960fa97c6 not found: ID does not exist" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.797136 4715 scope.go:117] "RemoveContainer" containerID="0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1" Dec 03 22:18:17 crc kubenswrapper[4715]: E1203 22:18:17.797440 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1\": container with ID starting with 0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1 not found: ID does not exist" containerID="0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1" Dec 03 22:18:17 crc kubenswrapper[4715]: I1203 22:18:17.797460 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1"} err="failed to get container status \"0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1\": rpc error: code = NotFound desc = could not find container \"0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1\": container with ID starting with 0472675bbbb3dd9320c714e8f63b3df9bdb5977f3d0990699e9e65d1748940a1 not found: ID does not exist" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.025990 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m68zt"] Dec 03 22:18:18 crc kubenswrapper[4715]: E1203 22:18:18.026381 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerName="extract-content" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.026397 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerName="extract-content" Dec 03 22:18:18 crc kubenswrapper[4715]: E1203 22:18:18.026409 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerName="extract-utilities" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.026415 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerName="extract-utilities" Dec 03 22:18:18 crc kubenswrapper[4715]: E1203 22:18:18.026454 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerName="registry-server" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.026459 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerName="registry-server" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.026636 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" containerName="registry-server" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.027257 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.031409 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.031674 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.031805 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-h88mg" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.037168 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m68zt"] Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.173533 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.173820 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-config-data\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.173848 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-scripts\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.173905 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp662\" (UniqueName: \"kubernetes.io/projected/008014fe-5f97-4de4-a7ef-4bf119aed38f-kube-api-access-hp662\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.276397 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp662\" (UniqueName: \"kubernetes.io/projected/008014fe-5f97-4de4-a7ef-4bf119aed38f-kube-api-access-hp662\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.276566 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.276615 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-config-data\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.276642 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-scripts\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.281603 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.283418 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-config-data\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.285453 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-scripts\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.293594 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp662\" (UniqueName: \"kubernetes.io/projected/008014fe-5f97-4de4-a7ef-4bf119aed38f-kube-api-access-hp662\") pod \"nova-cell0-conductor-db-sync-m68zt\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.345433 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.684589 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerStarted","Data":"908e708c6c5fc23797c8211f8a985125b4c2c9cb279733d4f01ece24fd68aa02"} Dec 03 22:18:18 crc kubenswrapper[4715]: I1203 22:18:18.782075 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m68zt"] Dec 03 22:18:18 crc kubenswrapper[4715]: W1203 22:18:18.782418 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice/crio-2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f WatchSource:0}: Error finding container 2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f: Status 404 returned error can't find the container with id 2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.344708 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.345041 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.393042 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.418040 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.644841 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fe015d5-6916-4c9a-a20f-d8be799a0e4c" path="/var/lib/kubelet/pods/9fe015d5-6916-4c9a-a20f-d8be799a0e4c/volumes" Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.694625 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-m68zt" event={"ID":"008014fe-5f97-4de4-a7ef-4bf119aed38f","Type":"ContainerStarted","Data":"2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f"} Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.698612 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerStarted","Data":"d0a23011b1f842cfc2046052210e74f89804bbd6c5d4820a0d356e78885b2ba4"} Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.698929 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:19 crc kubenswrapper[4715]: I1203 22:18:19.698971 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:21 crc kubenswrapper[4715]: I1203 22:18:21.476772 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:21 crc kubenswrapper[4715]: I1203 22:18:21.529366 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 22:18:21 crc kubenswrapper[4715]: I1203 22:18:21.723602 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerStarted","Data":"873128ba8afbcf5cf6fa4ea597295c358d4d925362dcd58c8a0a537448d6e555"} Dec 03 22:18:21 crc kubenswrapper[4715]: I1203 22:18:21.723981 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 22:18:21 crc kubenswrapper[4715]: I1203 22:18:21.754094 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.592662202 podStartE2EDuration="7.754072389s" podCreationTimestamp="2025-12-03 22:18:14 +0000 UTC" firstStartedPulling="2025-12-03 22:18:15.878069779 +0000 UTC m=+2192.620780374" lastFinishedPulling="2025-12-03 22:18:20.039479966 +0000 UTC m=+2196.782190561" observedRunningTime="2025-12-03 22:18:21.740420544 +0000 UTC m=+2198.483131159" watchObservedRunningTime="2025-12-03 22:18:21.754072389 +0000 UTC m=+2198.496782994" Dec 03 22:18:27 crc kubenswrapper[4715]: I1203 22:18:27.790005 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-m68zt" event={"ID":"008014fe-5f97-4de4-a7ef-4bf119aed38f","Type":"ContainerStarted","Data":"692ba571f2899b4dfe7b0e82b9945b9d217970de042398d4c450448f61b5c0dd"} Dec 03 22:18:27 crc kubenswrapper[4715]: I1203 22:18:27.833104 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-m68zt" podStartSLOduration=1.173016483 podStartE2EDuration="9.833086666s" podCreationTimestamp="2025-12-03 22:18:18 +0000 UTC" firstStartedPulling="2025-12-03 22:18:18.784142855 +0000 UTC m=+2195.526853450" lastFinishedPulling="2025-12-03 22:18:27.444212998 +0000 UTC m=+2204.186923633" observedRunningTime="2025-12-03 22:18:27.825559155 +0000 UTC m=+2204.568269770" watchObservedRunningTime="2025-12-03 22:18:27.833086666 +0000 UTC m=+2204.575797261" Dec 03 22:18:35 crc kubenswrapper[4715]: I1203 22:18:35.159946 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:18:35 crc kubenswrapper[4715]: I1203 22:18:35.161862 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:18:41 crc kubenswrapper[4715]: I1203 22:18:41.948260 4715 generic.go:334] "Generic (PLEG): container finished" podID="008014fe-5f97-4de4-a7ef-4bf119aed38f" containerID="692ba571f2899b4dfe7b0e82b9945b9d217970de042398d4c450448f61b5c0dd" exitCode=0 Dec 03 22:18:41 crc kubenswrapper[4715]: I1203 22:18:41.948366 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-m68zt" event={"ID":"008014fe-5f97-4de4-a7ef-4bf119aed38f","Type":"ContainerDied","Data":"692ba571f2899b4dfe7b0e82b9945b9d217970de042398d4c450448f61b5c0dd"} Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.414615 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.492052 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-config-data\") pod \"008014fe-5f97-4de4-a7ef-4bf119aed38f\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.492348 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-combined-ca-bundle\") pod \"008014fe-5f97-4de4-a7ef-4bf119aed38f\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.492442 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp662\" (UniqueName: \"kubernetes.io/projected/008014fe-5f97-4de4-a7ef-4bf119aed38f-kube-api-access-hp662\") pod \"008014fe-5f97-4de4-a7ef-4bf119aed38f\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.492667 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-scripts\") pod \"008014fe-5f97-4de4-a7ef-4bf119aed38f\" (UID: \"008014fe-5f97-4de4-a7ef-4bf119aed38f\") " Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.503393 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-scripts" (OuterVolumeSpecName: "scripts") pod "008014fe-5f97-4de4-a7ef-4bf119aed38f" (UID: "008014fe-5f97-4de4-a7ef-4bf119aed38f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.505909 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/008014fe-5f97-4de4-a7ef-4bf119aed38f-kube-api-access-hp662" (OuterVolumeSpecName: "kube-api-access-hp662") pod "008014fe-5f97-4de4-a7ef-4bf119aed38f" (UID: "008014fe-5f97-4de4-a7ef-4bf119aed38f"). InnerVolumeSpecName "kube-api-access-hp662". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.537361 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "008014fe-5f97-4de4-a7ef-4bf119aed38f" (UID: "008014fe-5f97-4de4-a7ef-4bf119aed38f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.547055 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-config-data" (OuterVolumeSpecName: "config-data") pod "008014fe-5f97-4de4-a7ef-4bf119aed38f" (UID: "008014fe-5f97-4de4-a7ef-4bf119aed38f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.596034 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.596086 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.596110 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp662\" (UniqueName: \"kubernetes.io/projected/008014fe-5f97-4de4-a7ef-4bf119aed38f-kube-api-access-hp662\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.596128 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008014fe-5f97-4de4-a7ef-4bf119aed38f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.994708 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-m68zt" event={"ID":"008014fe-5f97-4de4-a7ef-4bf119aed38f","Type":"ContainerDied","Data":"2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f"} Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.995078 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f" Dec 03 22:18:43 crc kubenswrapper[4715]: I1203 22:18:43.994939 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-m68zt" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.117064 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 22:18:44 crc kubenswrapper[4715]: E1203 22:18:44.117901 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="008014fe-5f97-4de4-a7ef-4bf119aed38f" containerName="nova-cell0-conductor-db-sync" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.117927 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="008014fe-5f97-4de4-a7ef-4bf119aed38f" containerName="nova-cell0-conductor-db-sync" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.118242 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="008014fe-5f97-4de4-a7ef-4bf119aed38f" containerName="nova-cell0-conductor-db-sync" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.119018 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.121786 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-h88mg" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.122004 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.129728 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.219552 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncjmd\" (UniqueName: \"kubernetes.io/projected/c328932f-799e-4f62-88c7-34c17ab3c0bc-kube-api-access-ncjmd\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.219640 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c328932f-799e-4f62-88c7-34c17ab3c0bc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.219796 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c328932f-799e-4f62-88c7-34c17ab3c0bc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: E1203 22:18:44.293087 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice/crio-2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f\": RecentStats: unable to find data in memory cache]" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.321426 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncjmd\" (UniqueName: \"kubernetes.io/projected/c328932f-799e-4f62-88c7-34c17ab3c0bc-kube-api-access-ncjmd\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.321547 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c328932f-799e-4f62-88c7-34c17ab3c0bc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.321625 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c328932f-799e-4f62-88c7-34c17ab3c0bc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.327168 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c328932f-799e-4f62-88c7-34c17ab3c0bc-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.327440 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c328932f-799e-4f62-88c7-34c17ab3c0bc-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.344258 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncjmd\" (UniqueName: \"kubernetes.io/projected/c328932f-799e-4f62-88c7-34c17ab3c0bc-kube-api-access-ncjmd\") pod \"nova-cell0-conductor-0\" (UID: \"c328932f-799e-4f62-88c7-34c17ab3c0bc\") " pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.440589 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:44 crc kubenswrapper[4715]: I1203 22:18:44.957149 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 22:18:45 crc kubenswrapper[4715]: I1203 22:18:45.009674 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c328932f-799e-4f62-88c7-34c17ab3c0bc","Type":"ContainerStarted","Data":"7842705cce583bf879aafc28f00d36221ceb1809912caac5a097eaceb9166822"} Dec 03 22:18:45 crc kubenswrapper[4715]: I1203 22:18:45.368919 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 22:18:47 crc kubenswrapper[4715]: I1203 22:18:47.037576 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c328932f-799e-4f62-88c7-34c17ab3c0bc","Type":"ContainerStarted","Data":"520a09f6442bca9ad3ce672056d3cccdecda7e7358ac21ac92a397fe01558d52"} Dec 03 22:18:47 crc kubenswrapper[4715]: I1203 22:18:47.037857 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:47 crc kubenswrapper[4715]: I1203 22:18:47.081765 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.081738967 podStartE2EDuration="3.081738967s" podCreationTimestamp="2025-12-03 22:18:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:18:47.06462446 +0000 UTC m=+2223.807335065" watchObservedRunningTime="2025-12-03 22:18:47.081738967 +0000 UTC m=+2223.824449602" Dec 03 22:18:49 crc kubenswrapper[4715]: I1203 22:18:49.196969 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:18:49 crc kubenswrapper[4715]: I1203 22:18:49.198492 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="e514cbad-5355-479e-be39-a625b874551c" containerName="kube-state-metrics" containerID="cri-o://38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f" gracePeriod=30 Dec 03 22:18:49 crc kubenswrapper[4715]: I1203 22:18:49.616561 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 22:18:49 crc kubenswrapper[4715]: I1203 22:18:49.736787 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnp5f\" (UniqueName: \"kubernetes.io/projected/e514cbad-5355-479e-be39-a625b874551c-kube-api-access-fnp5f\") pod \"e514cbad-5355-479e-be39-a625b874551c\" (UID: \"e514cbad-5355-479e-be39-a625b874551c\") " Dec 03 22:18:49 crc kubenswrapper[4715]: I1203 22:18:49.744166 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e514cbad-5355-479e-be39-a625b874551c-kube-api-access-fnp5f" (OuterVolumeSpecName: "kube-api-access-fnp5f") pod "e514cbad-5355-479e-be39-a625b874551c" (UID: "e514cbad-5355-479e-be39-a625b874551c"). InnerVolumeSpecName "kube-api-access-fnp5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:18:49 crc kubenswrapper[4715]: I1203 22:18:49.839936 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnp5f\" (UniqueName: \"kubernetes.io/projected/e514cbad-5355-479e-be39-a625b874551c-kube-api-access-fnp5f\") on node \"crc\" DevicePath \"\"" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.071927 4715 generic.go:334] "Generic (PLEG): container finished" podID="e514cbad-5355-479e-be39-a625b874551c" containerID="38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f" exitCode=2 Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.071989 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.072007 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e514cbad-5355-479e-be39-a625b874551c","Type":"ContainerDied","Data":"38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f"} Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.072069 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"e514cbad-5355-479e-be39-a625b874551c","Type":"ContainerDied","Data":"5f134ca1511de95331999f4b8b6411c777749f153cf5ef534a6c792cac9eeadf"} Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.072103 4715 scope.go:117] "RemoveContainer" containerID="38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.109285 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.111426 4715 scope.go:117] "RemoveContainer" containerID="38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f" Dec 03 22:18:50 crc kubenswrapper[4715]: E1203 22:18:50.112121 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f\": container with ID starting with 38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f not found: ID does not exist" containerID="38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.112174 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f"} err="failed to get container status \"38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f\": rpc error: code = NotFound desc = could not find container \"38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f\": container with ID starting with 38f40a235ae478ab0d352e4b5e4792c82db566bf309318ba020e929976236a8f not found: ID does not exist" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.119119 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.135229 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:18:50 crc kubenswrapper[4715]: E1203 22:18:50.135989 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e514cbad-5355-479e-be39-a625b874551c" containerName="kube-state-metrics" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.136027 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e514cbad-5355-479e-be39-a625b874551c" containerName="kube-state-metrics" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.136463 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e514cbad-5355-479e-be39-a625b874551c" containerName="kube-state-metrics" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.137710 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.139461 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.145012 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.152282 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.247305 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxz64\" (UniqueName: \"kubernetes.io/projected/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-api-access-zxz64\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.247434 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.247537 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.247710 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.350166 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxz64\" (UniqueName: \"kubernetes.io/projected/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-api-access-zxz64\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.350656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.350754 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.350798 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.355464 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.355892 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.357871 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.377395 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxz64\" (UniqueName: \"kubernetes.io/projected/11b5a8ae-b831-453c-964b-1631e5d1280c-kube-api-access-zxz64\") pod \"kube-state-metrics-0\" (UID: \"11b5a8ae-b831-453c-964b-1631e5d1280c\") " pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.504462 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.952194 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.952859 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="ceilometer-central-agent" containerID="cri-o://2298e6b06d0ef1f39ff8e2d48041a92c76c9b3b2c5632e91a003824acefbc015" gracePeriod=30 Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.952990 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="ceilometer-notification-agent" containerID="cri-o://908e708c6c5fc23797c8211f8a985125b4c2c9cb279733d4f01ece24fd68aa02" gracePeriod=30 Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.952928 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="proxy-httpd" containerID="cri-o://873128ba8afbcf5cf6fa4ea597295c358d4d925362dcd58c8a0a537448d6e555" gracePeriod=30 Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.952914 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="sg-core" containerID="cri-o://d0a23011b1f842cfc2046052210e74f89804bbd6c5d4820a0d356e78885b2ba4" gracePeriod=30 Dec 03 22:18:50 crc kubenswrapper[4715]: I1203 22:18:50.983824 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 22:18:51 crc kubenswrapper[4715]: I1203 22:18:51.086529 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"11b5a8ae-b831-453c-964b-1631e5d1280c","Type":"ContainerStarted","Data":"f9be07e104e46a50d6070a747bb4dbf9d7362c8d899b95c53d28e18e40101ec1"} Dec 03 22:18:51 crc kubenswrapper[4715]: I1203 22:18:51.644880 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e514cbad-5355-479e-be39-a625b874551c" path="/var/lib/kubelet/pods/e514cbad-5355-479e-be39-a625b874551c/volumes" Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.102285 4715 generic.go:334] "Generic (PLEG): container finished" podID="d226be51-32f8-4710-af9f-afc1253b1485" containerID="873128ba8afbcf5cf6fa4ea597295c358d4d925362dcd58c8a0a537448d6e555" exitCode=0 Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.102337 4715 generic.go:334] "Generic (PLEG): container finished" podID="d226be51-32f8-4710-af9f-afc1253b1485" containerID="d0a23011b1f842cfc2046052210e74f89804bbd6c5d4820a0d356e78885b2ba4" exitCode=2 Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.102351 4715 generic.go:334] "Generic (PLEG): container finished" podID="d226be51-32f8-4710-af9f-afc1253b1485" containerID="2298e6b06d0ef1f39ff8e2d48041a92c76c9b3b2c5632e91a003824acefbc015" exitCode=0 Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.102392 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerDied","Data":"873128ba8afbcf5cf6fa4ea597295c358d4d925362dcd58c8a0a537448d6e555"} Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.102458 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerDied","Data":"d0a23011b1f842cfc2046052210e74f89804bbd6c5d4820a0d356e78885b2ba4"} Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.102476 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerDied","Data":"2298e6b06d0ef1f39ff8e2d48041a92c76c9b3b2c5632e91a003824acefbc015"} Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.104716 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"11b5a8ae-b831-453c-964b-1631e5d1280c","Type":"ContainerStarted","Data":"39f8076120e37debe2a655145721eb463ce0af07e876d6280a77050b8b75769e"} Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.104880 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 22:18:52 crc kubenswrapper[4715]: I1203 22:18:52.139469 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.571401742 podStartE2EDuration="2.139438493s" podCreationTimestamp="2025-12-03 22:18:50 +0000 UTC" firstStartedPulling="2025-12-03 22:18:51.001985525 +0000 UTC m=+2227.744696120" lastFinishedPulling="2025-12-03 22:18:51.570022266 +0000 UTC m=+2228.312732871" observedRunningTime="2025-12-03 22:18:52.124193255 +0000 UTC m=+2228.866903850" watchObservedRunningTime="2025-12-03 22:18:52.139438493 +0000 UTC m=+2228.882149128" Dec 03 22:18:54 crc kubenswrapper[4715]: I1203 22:18:54.483314 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 22:18:54 crc kubenswrapper[4715]: E1203 22:18:54.549293 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice/crio-2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice\": RecentStats: unable to find data in memory cache]" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.004898 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-jkd9h"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.006482 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.012242 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.012431 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.028974 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jkd9h"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.156641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5g9x\" (UniqueName: \"kubernetes.io/projected/1ac13b78-4213-452d-8619-9a2bc7ee5338-kube-api-access-g5g9x\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.156926 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-config-data\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.157031 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-scripts\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.157144 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.182012 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.185494 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.191094 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.196318 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.197463 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.199173 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.209787 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.218446 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.259772 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5g9x\" (UniqueName: \"kubernetes.io/projected/1ac13b78-4213-452d-8619-9a2bc7ee5338-kube-api-access-g5g9x\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.259832 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.259882 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-config-data\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.259902 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-config-data\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.259937 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-scripts\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.259964 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkk8q\" (UniqueName: \"kubernetes.io/projected/5d7be858-2aa9-45ff-b195-36738f727990-kube-api-access-pkk8q\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.259993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.260035 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7be858-2aa9-45ff-b195-36738f727990-logs\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.269357 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-scripts\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.271904 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.280229 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5g9x\" (UniqueName: \"kubernetes.io/projected/1ac13b78-4213-452d-8619-9a2bc7ee5338-kube-api-access-g5g9x\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.280350 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-config-data\") pod \"nova-cell0-cell-mapping-jkd9h\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.322136 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.361725 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.361794 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7be858-2aa9-45ff-b195-36738f727990-logs\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.361961 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.362048 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck9hn\" (UniqueName: \"kubernetes.io/projected/f199e727-bde9-4b7f-9c09-e5b17a13beda-kube-api-access-ck9hn\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.362084 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-config-data\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.362210 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-config-data\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.362350 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7be858-2aa9-45ff-b195-36738f727990-logs\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.362411 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkk8q\" (UniqueName: \"kubernetes.io/projected/5d7be858-2aa9-45ff-b195-36738f727990-kube-api-access-pkk8q\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.366151 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-config-data\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.404424 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkk8q\" (UniqueName: \"kubernetes.io/projected/5d7be858-2aa9-45ff-b195-36738f727990-kube-api-access-pkk8q\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.416421 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.416770 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.418317 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.429974 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.435727 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.455082 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.457661 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.460385 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.464968 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck9hn\" (UniqueName: \"kubernetes.io/projected/f199e727-bde9-4b7f-9c09-e5b17a13beda-kube-api-access-ck9hn\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.465023 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-config-data\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.465128 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.485336 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-config-data\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.485811 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.488565 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.491441 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck9hn\" (UniqueName: \"kubernetes.io/projected/f199e727-bde9-4b7f-9c09-e5b17a13beda-kube-api-access-ck9hn\") pod \"nova-scheduler-0\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.506630 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.518081 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.566342 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-6b88c"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.607114 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.607167 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.607303 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.607393 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v999w\" (UniqueName: \"kubernetes.io/projected/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-kube-api-access-v999w\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.607417 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-logs\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.607531 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-config-data\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.607564 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdpdv\" (UniqueName: \"kubernetes.io/projected/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-kube-api-access-hdpdv\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.610303 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.625589 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-6b88c"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.718552 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-config-data\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.718796 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdpdv\" (UniqueName: \"kubernetes.io/projected/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-kube-api-access-hdpdv\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.718839 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-svc\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.718896 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.718915 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.718937 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.718955 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-config\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.718980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.719019 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.719042 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.719060 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbgt7\" (UniqueName: \"kubernetes.io/projected/b3f721fb-a909-4614-8f12-ae64db2c78c9-kube-api-access-vbgt7\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.719088 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v999w\" (UniqueName: \"kubernetes.io/projected/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-kube-api-access-v999w\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.719104 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-logs\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.719989 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-logs\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.724936 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.725381 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-config-data\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.727264 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.727327 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.750869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v999w\" (UniqueName: \"kubernetes.io/projected/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-kube-api-access-v999w\") pod \"nova-metadata-0\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.757365 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdpdv\" (UniqueName: \"kubernetes.io/projected/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-kube-api-access-hdpdv\") pod \"nova-cell1-novncproxy-0\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.820445 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-svc\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.820561 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.820581 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-config\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.820619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.820665 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.820686 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbgt7\" (UniqueName: \"kubernetes.io/projected/b3f721fb-a909-4614-8f12-ae64db2c78c9-kube-api-access-vbgt7\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.821722 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-svc\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.822052 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-config\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.822357 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.822888 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.822929 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.838054 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbgt7\" (UniqueName: \"kubernetes.io/projected/b3f721fb-a909-4614-8f12-ae64db2c78c9-kube-api-access-vbgt7\") pod \"dnsmasq-dns-bccf8f775-6b88c\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.855914 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.914429 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.942094 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.981806 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-svdtv"] Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.983298 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.992000 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 22:18:55 crc kubenswrapper[4715]: I1203 22:18:55.992031 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.020008 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jkd9h"] Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.043815 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-svdtv"] Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.129756 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqpdd\" (UniqueName: \"kubernetes.io/projected/f05149da-e2c3-4bbb-9b7b-e51f6d022875-kube-api-access-sqpdd\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.129844 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-config-data\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.129896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-scripts\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.129922 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.159016 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jkd9h" event={"ID":"1ac13b78-4213-452d-8619-9a2bc7ee5338","Type":"ContainerStarted","Data":"1fd7de3821eed11901198f7b2ba96342bba838a7e061dea779446f007d7ea99a"} Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.164575 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.178712 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.231865 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-config-data\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.232982 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-scripts\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.233035 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.233298 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqpdd\" (UniqueName: \"kubernetes.io/projected/f05149da-e2c3-4bbb-9b7b-e51f6d022875-kube-api-access-sqpdd\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.235758 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-scripts\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.243340 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-config-data\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.244683 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.250525 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqpdd\" (UniqueName: \"kubernetes.io/projected/f05149da-e2c3-4bbb-9b7b-e51f6d022875-kube-api-access-sqpdd\") pod \"nova-cell1-conductor-db-sync-svdtv\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.315393 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.447586 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.580756 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-6b88c"] Dec 03 22:18:56 crc kubenswrapper[4715]: W1203 22:18:56.584897 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3f721fb_a909_4614_8f12_ae64db2c78c9.slice/crio-785d959dd2c30c9c023a1d4d3ec32e64c20668144df0af2bbe636223d8019d61 WatchSource:0}: Error finding container 785d959dd2c30c9c023a1d4d3ec32e64c20668144df0af2bbe636223d8019d61: Status 404 returned error can't find the container with id 785d959dd2c30c9c023a1d4d3ec32e64c20668144df0af2bbe636223d8019d61 Dec 03 22:18:56 crc kubenswrapper[4715]: W1203 22:18:56.586214 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78f4d687_a0a9_49ef_86a6_4e9ebf6af0f8.slice/crio-4a7aadbfd0f4303461a5231d71ec42aadf85a4ccc558ae0480b4dc38cbfb4dc7 WatchSource:0}: Error finding container 4a7aadbfd0f4303461a5231d71ec42aadf85a4ccc558ae0480b4dc38cbfb4dc7: Status 404 returned error can't find the container with id 4a7aadbfd0f4303461a5231d71ec42aadf85a4ccc558ae0480b4dc38cbfb4dc7 Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.602723 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:18:56 crc kubenswrapper[4715]: W1203 22:18:56.800739 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf05149da_e2c3_4bbb_9b7b_e51f6d022875.slice/crio-afb6f288619a16001376d7932b860c58dc1a1f8417edc57c3f78d64800905012 WatchSource:0}: Error finding container afb6f288619a16001376d7932b860c58dc1a1f8417edc57c3f78d64800905012: Status 404 returned error can't find the container with id afb6f288619a16001376d7932b860c58dc1a1f8417edc57c3f78d64800905012 Dec 03 22:18:56 crc kubenswrapper[4715]: I1203 22:18:56.802096 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-svdtv"] Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.179589 4715 generic.go:334] "Generic (PLEG): container finished" podID="b3f721fb-a909-4614-8f12-ae64db2c78c9" containerID="f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d" exitCode=0 Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.179673 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" event={"ID":"b3f721fb-a909-4614-8f12-ae64db2c78c9","Type":"ContainerDied","Data":"f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.179710 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" event={"ID":"b3f721fb-a909-4614-8f12-ae64db2c78c9","Type":"ContainerStarted","Data":"785d959dd2c30c9c023a1d4d3ec32e64c20668144df0af2bbe636223d8019d61"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.183214 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f199e727-bde9-4b7f-9c09-e5b17a13beda","Type":"ContainerStarted","Data":"993f96d33d12b942b4c672d496922a3be91e23dd59963871dd4c3ac1fbf060d4"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.188395 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jkd9h" event={"ID":"1ac13b78-4213-452d-8619-9a2bc7ee5338","Type":"ContainerStarted","Data":"9260bbec95f69642bf1876790b1f2325d6377636275474c7052e74f51c546005"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.191188 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-svdtv" event={"ID":"f05149da-e2c3-4bbb-9b7b-e51f6d022875","Type":"ContainerStarted","Data":"d38a7035297765d245716784ed4f49f1e33b03bdd2f0cd344fa42d5fddd3f62a"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.191228 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-svdtv" event={"ID":"f05149da-e2c3-4bbb-9b7b-e51f6d022875","Type":"ContainerStarted","Data":"afb6f288619a16001376d7932b860c58dc1a1f8417edc57c3f78d64800905012"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.192596 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5d7be858-2aa9-45ff-b195-36738f727990","Type":"ContainerStarted","Data":"32feb3e913fa766ae016ef277721d615e27a9171666003db6977f3df2157bfc8"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.206279 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8","Type":"ContainerStarted","Data":"4a7aadbfd0f4303461a5231d71ec42aadf85a4ccc558ae0480b4dc38cbfb4dc7"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.213546 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63","Type":"ContainerStarted","Data":"38cc7a0eb158b91e3be83a08e5e66f8c93e42ed4a6cb744e1813db66828afdf8"} Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.222829 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-jkd9h" podStartSLOduration=3.222813916 podStartE2EDuration="3.222813916s" podCreationTimestamp="2025-12-03 22:18:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:18:57.217631327 +0000 UTC m=+2233.960341942" watchObservedRunningTime="2025-12-03 22:18:57.222813916 +0000 UTC m=+2233.965524511" Dec 03 22:18:57 crc kubenswrapper[4715]: I1203 22:18:57.242864 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-svdtv" podStartSLOduration=2.242842841 podStartE2EDuration="2.242842841s" podCreationTimestamp="2025-12-03 22:18:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:18:57.235380642 +0000 UTC m=+2233.978091237" watchObservedRunningTime="2025-12-03 22:18:57.242842841 +0000 UTC m=+2233.985553436" Dec 03 22:18:58 crc kubenswrapper[4715]: I1203 22:18:58.229607 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" event={"ID":"b3f721fb-a909-4614-8f12-ae64db2c78c9","Type":"ContainerStarted","Data":"6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd"} Dec 03 22:18:59 crc kubenswrapper[4715]: I1203 22:18:59.473351 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:18:59 crc kubenswrapper[4715]: I1203 22:18:59.484747 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:19:00 crc kubenswrapper[4715]: I1203 22:19:00.703030 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.271433 4715 generic.go:334] "Generic (PLEG): container finished" podID="d226be51-32f8-4710-af9f-afc1253b1485" containerID="908e708c6c5fc23797c8211f8a985125b4c2c9cb279733d4f01ece24fd68aa02" exitCode=0 Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.271883 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerDied","Data":"908e708c6c5fc23797c8211f8a985125b4c2c9cb279733d4f01ece24fd68aa02"} Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.271958 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.300495 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" podStartSLOduration=6.300467102 podStartE2EDuration="6.300467102s" podCreationTimestamp="2025-12-03 22:18:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:01.287810644 +0000 UTC m=+2238.030521239" watchObservedRunningTime="2025-12-03 22:19:01.300467102 +0000 UTC m=+2238.043177697" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.661429 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.757768 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-run-httpd\") pod \"d226be51-32f8-4710-af9f-afc1253b1485\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.758250 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnjfj\" (UniqueName: \"kubernetes.io/projected/d226be51-32f8-4710-af9f-afc1253b1485-kube-api-access-vnjfj\") pod \"d226be51-32f8-4710-af9f-afc1253b1485\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.758319 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-scripts\") pod \"d226be51-32f8-4710-af9f-afc1253b1485\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.758355 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-combined-ca-bundle\") pod \"d226be51-32f8-4710-af9f-afc1253b1485\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.758195 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d226be51-32f8-4710-af9f-afc1253b1485" (UID: "d226be51-32f8-4710-af9f-afc1253b1485"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.759451 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-config-data\") pod \"d226be51-32f8-4710-af9f-afc1253b1485\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.759559 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-sg-core-conf-yaml\") pod \"d226be51-32f8-4710-af9f-afc1253b1485\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.759587 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-log-httpd\") pod \"d226be51-32f8-4710-af9f-afc1253b1485\" (UID: \"d226be51-32f8-4710-af9f-afc1253b1485\") " Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.760666 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.762605 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d226be51-32f8-4710-af9f-afc1253b1485" (UID: "d226be51-32f8-4710-af9f-afc1253b1485"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.765486 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d226be51-32f8-4710-af9f-afc1253b1485-kube-api-access-vnjfj" (OuterVolumeSpecName: "kube-api-access-vnjfj") pod "d226be51-32f8-4710-af9f-afc1253b1485" (UID: "d226be51-32f8-4710-af9f-afc1253b1485"). InnerVolumeSpecName "kube-api-access-vnjfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.790795 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-scripts" (OuterVolumeSpecName: "scripts") pod "d226be51-32f8-4710-af9f-afc1253b1485" (UID: "d226be51-32f8-4710-af9f-afc1253b1485"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.795759 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d226be51-32f8-4710-af9f-afc1253b1485" (UID: "d226be51-32f8-4710-af9f-afc1253b1485"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.841880 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d226be51-32f8-4710-af9f-afc1253b1485" (UID: "d226be51-32f8-4710-af9f-afc1253b1485"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.863305 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.863338 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.863352 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.863363 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d226be51-32f8-4710-af9f-afc1253b1485-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.863376 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnjfj\" (UniqueName: \"kubernetes.io/projected/d226be51-32f8-4710-af9f-afc1253b1485-kube-api-access-vnjfj\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.877312 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-config-data" (OuterVolumeSpecName: "config-data") pod "d226be51-32f8-4710-af9f-afc1253b1485" (UID: "d226be51-32f8-4710-af9f-afc1253b1485"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:01 crc kubenswrapper[4715]: I1203 22:19:01.965607 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d226be51-32f8-4710-af9f-afc1253b1485-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.289048 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.289212 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d226be51-32f8-4710-af9f-afc1253b1485","Type":"ContainerDied","Data":"2e923eadba35f98115a5b7508189cfc0a073013cec81feb42b049d7b904e8822"} Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.289388 4715 scope.go:117] "RemoveContainer" containerID="873128ba8afbcf5cf6fa4ea597295c358d4d925362dcd58c8a0a537448d6e555" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.352797 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.369360 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.388692 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:02 crc kubenswrapper[4715]: E1203 22:19:02.389095 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="sg-core" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.389116 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="sg-core" Dec 03 22:19:02 crc kubenswrapper[4715]: E1203 22:19:02.389140 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="ceilometer-notification-agent" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.389146 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="ceilometer-notification-agent" Dec 03 22:19:02 crc kubenswrapper[4715]: E1203 22:19:02.389158 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="proxy-httpd" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.389164 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="proxy-httpd" Dec 03 22:19:02 crc kubenswrapper[4715]: E1203 22:19:02.389184 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="ceilometer-central-agent" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.389190 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="ceilometer-central-agent" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.389375 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="sg-core" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.389396 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="proxy-httpd" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.389409 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="ceilometer-central-agent" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.389424 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d226be51-32f8-4710-af9f-afc1253b1485" containerName="ceilometer-notification-agent" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.391889 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.394053 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.394410 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.400329 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.403079 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.578639 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wssbf\" (UniqueName: \"kubernetes.io/projected/f582783f-a94f-4910-962b-51f24976b3e6-kube-api-access-wssbf\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.578716 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-log-httpd\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.578752 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.578774 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.578816 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-config-data\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.578875 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-scripts\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.578899 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.578933 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-run-httpd\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.682329 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wssbf\" (UniqueName: \"kubernetes.io/projected/f582783f-a94f-4910-962b-51f24976b3e6-kube-api-access-wssbf\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.682419 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-log-httpd\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.682460 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.682483 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.682554 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-config-data\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.682644 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-scripts\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.682677 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.682726 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-run-httpd\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.683244 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-run-httpd\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.683815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-log-httpd\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.688798 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.689645 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-config-data\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.691232 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-scripts\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.691902 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.693043 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.704556 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wssbf\" (UniqueName: \"kubernetes.io/projected/f582783f-a94f-4910-962b-51f24976b3e6-kube-api-access-wssbf\") pod \"ceilometer-0\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " pod="openstack/ceilometer-0" Dec 03 22:19:02 crc kubenswrapper[4715]: I1203 22:19:02.718672 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:03 crc kubenswrapper[4715]: I1203 22:19:03.359081 4715 scope.go:117] "RemoveContainer" containerID="d0a23011b1f842cfc2046052210e74f89804bbd6c5d4820a0d356e78885b2ba4" Dec 03 22:19:03 crc kubenswrapper[4715]: I1203 22:19:03.427172 4715 scope.go:117] "RemoveContainer" containerID="908e708c6c5fc23797c8211f8a985125b4c2c9cb279733d4f01ece24fd68aa02" Dec 03 22:19:03 crc kubenswrapper[4715]: I1203 22:19:03.465474 4715 scope.go:117] "RemoveContainer" containerID="2298e6b06d0ef1f39ff8e2d48041a92c76c9b3b2c5632e91a003824acefbc015" Dec 03 22:19:03 crc kubenswrapper[4715]: I1203 22:19:03.647639 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d226be51-32f8-4710-af9f-afc1253b1485" path="/var/lib/kubelet/pods/d226be51-32f8-4710-af9f-afc1253b1485/volumes" Dec 03 22:19:03 crc kubenswrapper[4715]: I1203 22:19:03.941218 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.310219 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8","Type":"ContainerStarted","Data":"57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d"} Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.310326 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d" gracePeriod=30 Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.312751 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63","Type":"ContainerStarted","Data":"6c1d67be0f901af6d566732377d32d0ce6ce7f3ae85a61efffcf6c009a643293"} Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.315746 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f199e727-bde9-4b7f-9c09-e5b17a13beda","Type":"ContainerStarted","Data":"77231f7728a88f0edb7e29432c877cd91f0f46986c5b237eef2a47a80c138f42"} Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.320161 4715 generic.go:334] "Generic (PLEG): container finished" podID="1ac13b78-4213-452d-8619-9a2bc7ee5338" containerID="9260bbec95f69642bf1876790b1f2325d6377636275474c7052e74f51c546005" exitCode=0 Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.320189 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jkd9h" event={"ID":"1ac13b78-4213-452d-8619-9a2bc7ee5338","Type":"ContainerDied","Data":"9260bbec95f69642bf1876790b1f2325d6377636275474c7052e74f51c546005"} Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.321848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerStarted","Data":"f73f0918d87642ae79cd9cbc9f4f3ea58d09f028f416d3995a4d531825df670f"} Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.328323 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5d7be858-2aa9-45ff-b195-36738f727990","Type":"ContainerStarted","Data":"2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92"} Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.338659 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.482042238 podStartE2EDuration="9.338644801s" podCreationTimestamp="2025-12-03 22:18:55 +0000 UTC" firstStartedPulling="2025-12-03 22:18:56.588126643 +0000 UTC m=+2233.330837238" lastFinishedPulling="2025-12-03 22:19:03.444729196 +0000 UTC m=+2240.187439801" observedRunningTime="2025-12-03 22:19:04.334763927 +0000 UTC m=+2241.077474522" watchObservedRunningTime="2025-12-03 22:19:04.338644801 +0000 UTC m=+2241.081355396" Dec 03 22:19:04 crc kubenswrapper[4715]: I1203 22:19:04.370204 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.06028535 podStartE2EDuration="9.370173705s" podCreationTimestamp="2025-12-03 22:18:55 +0000 UTC" firstStartedPulling="2025-12-03 22:18:56.15802244 +0000 UTC m=+2232.900733035" lastFinishedPulling="2025-12-03 22:19:03.467910785 +0000 UTC m=+2240.210621390" observedRunningTime="2025-12-03 22:19:04.356251462 +0000 UTC m=+2241.098962057" watchObservedRunningTime="2025-12-03 22:19:04.370173705 +0000 UTC m=+2241.112884300" Dec 03 22:19:04 crc kubenswrapper[4715]: E1203 22:19:04.865918 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice/crio-2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f\": RecentStats: unable to find data in memory cache]" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.159750 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.160127 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.160180 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.160738 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.160800 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" gracePeriod=600 Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.341920 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63","Type":"ContainerStarted","Data":"4178b3536c867700181c222e6dcc90b6a769b5e2fb10972b1829e75b67a91c95"} Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.342109 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerName="nova-metadata-log" containerID="cri-o://6c1d67be0f901af6d566732377d32d0ce6ce7f3ae85a61efffcf6c009a643293" gracePeriod=30 Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.342696 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerName="nova-metadata-metadata" containerID="cri-o://4178b3536c867700181c222e6dcc90b6a769b5e2fb10972b1829e75b67a91c95" gracePeriod=30 Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.351096 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" exitCode=0 Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.351193 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc"} Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.351240 4715 scope.go:117] "RemoveContainer" containerID="2a7b507f93c05d0aab4e197e7602419073e36d87d0fa5a60853ef24a637d4567" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.354919 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerStarted","Data":"5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb"} Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.359749 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5d7be858-2aa9-45ff-b195-36738f727990","Type":"ContainerStarted","Data":"52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05"} Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.383719 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.402145403 podStartE2EDuration="10.383692178s" podCreationTimestamp="2025-12-03 22:18:55 +0000 UTC" firstStartedPulling="2025-12-03 22:18:56.463180141 +0000 UTC m=+2233.205890736" lastFinishedPulling="2025-12-03 22:19:03.444726916 +0000 UTC m=+2240.187437511" observedRunningTime="2025-12-03 22:19:05.372062677 +0000 UTC m=+2242.114773292" watchObservedRunningTime="2025-12-03 22:19:05.383692178 +0000 UTC m=+2242.126402773" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.401538 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.124823328 podStartE2EDuration="10.401520055s" podCreationTimestamp="2025-12-03 22:18:55 +0000 UTC" firstStartedPulling="2025-12-03 22:18:56.191198998 +0000 UTC m=+2232.933909593" lastFinishedPulling="2025-12-03 22:19:03.467895725 +0000 UTC m=+2240.210606320" observedRunningTime="2025-12-03 22:19:05.398087293 +0000 UTC m=+2242.140797888" watchObservedRunningTime="2025-12-03 22:19:05.401520055 +0000 UTC m=+2242.144230650" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.529299 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.529363 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.539907 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.540181 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.584020 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 22:19:05 crc kubenswrapper[4715]: E1203 22:19:05.844397 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.852300 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.856675 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.856745 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.914789 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.944702 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.975022 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5g9x\" (UniqueName: \"kubernetes.io/projected/1ac13b78-4213-452d-8619-9a2bc7ee5338-kube-api-access-g5g9x\") pod \"1ac13b78-4213-452d-8619-9a2bc7ee5338\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.975827 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-combined-ca-bundle\") pod \"1ac13b78-4213-452d-8619-9a2bc7ee5338\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.975925 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-config-data\") pod \"1ac13b78-4213-452d-8619-9a2bc7ee5338\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.975987 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-scripts\") pod \"1ac13b78-4213-452d-8619-9a2bc7ee5338\" (UID: \"1ac13b78-4213-452d-8619-9a2bc7ee5338\") " Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.978904 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ac13b78-4213-452d-8619-9a2bc7ee5338-kube-api-access-g5g9x" (OuterVolumeSpecName: "kube-api-access-g5g9x") pod "1ac13b78-4213-452d-8619-9a2bc7ee5338" (UID: "1ac13b78-4213-452d-8619-9a2bc7ee5338"). InnerVolumeSpecName "kube-api-access-g5g9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:05 crc kubenswrapper[4715]: I1203 22:19:05.984820 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-scripts" (OuterVolumeSpecName: "scripts") pod "1ac13b78-4213-452d-8619-9a2bc7ee5338" (UID: "1ac13b78-4213-452d-8619-9a2bc7ee5338"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.026491 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-fhfjs"] Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.026795 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" podUID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" containerName="dnsmasq-dns" containerID="cri-o://d1971a09d49c8130bc0a70f573319660f3e7e1990d43eafef19855a34f28f459" gracePeriod=10 Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.078680 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.078706 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5g9x\" (UniqueName: \"kubernetes.io/projected/1ac13b78-4213-452d-8619-9a2bc7ee5338-kube-api-access-g5g9x\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.084272 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ac13b78-4213-452d-8619-9a2bc7ee5338" (UID: "1ac13b78-4213-452d-8619-9a2bc7ee5338"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.106938 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-config-data" (OuterVolumeSpecName: "config-data") pod "1ac13b78-4213-452d-8619-9a2bc7ee5338" (UID: "1ac13b78-4213-452d-8619-9a2bc7ee5338"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.182727 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.183165 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ac13b78-4213-452d-8619-9a2bc7ee5338-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.386776 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerStarted","Data":"618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef"} Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.392122 4715 generic.go:334] "Generic (PLEG): container finished" podID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerID="4178b3536c867700181c222e6dcc90b6a769b5e2fb10972b1829e75b67a91c95" exitCode=0 Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.392174 4715 generic.go:334] "Generic (PLEG): container finished" podID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerID="6c1d67be0f901af6d566732377d32d0ce6ce7f3ae85a61efffcf6c009a643293" exitCode=143 Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.392233 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63","Type":"ContainerDied","Data":"4178b3536c867700181c222e6dcc90b6a769b5e2fb10972b1829e75b67a91c95"} Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.392273 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63","Type":"ContainerDied","Data":"6c1d67be0f901af6d566732377d32d0ce6ce7f3ae85a61efffcf6c009a643293"} Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.395213 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jkd9h" event={"ID":"1ac13b78-4213-452d-8619-9a2bc7ee5338","Type":"ContainerDied","Data":"1fd7de3821eed11901198f7b2ba96342bba838a7e061dea779446f007d7ea99a"} Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.395300 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fd7de3821eed11901198f7b2ba96342bba838a7e061dea779446f007d7ea99a" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.395442 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jkd9h" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.403168 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:19:06 crc kubenswrapper[4715]: E1203 22:19:06.403912 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.428241 4715 generic.go:334] "Generic (PLEG): container finished" podID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" containerID="d1971a09d49c8130bc0a70f573319660f3e7e1990d43eafef19855a34f28f459" exitCode=0 Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.429225 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" event={"ID":"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f","Type":"ContainerDied","Data":"d1971a09d49c8130bc0a70f573319660f3e7e1990d43eafef19855a34f28f459"} Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.450124 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.466208 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.616207 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-logs\") pod \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.616366 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v999w\" (UniqueName: \"kubernetes.io/projected/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-kube-api-access-v999w\") pod \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.616488 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-combined-ca-bundle\") pod \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.616608 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-config-data\") pod \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\" (UID: \"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.617173 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.617376 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.618008 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-logs" (OuterVolumeSpecName: "logs") pod "1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" (UID: "1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.629757 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-kube-api-access-v999w" (OuterVolumeSpecName: "kube-api-access-v999w") pod "1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" (UID: "1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63"). InnerVolumeSpecName "kube-api-access-v999w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.639007 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.674638 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-config-data" (OuterVolumeSpecName: "config-data") pod "1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" (UID: "1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.676922 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" (UID: "1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.724312 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.724520 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.724607 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v999w\" (UniqueName: \"kubernetes.io/projected/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-kube-api-access-v999w\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.724682 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.757376 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.826566 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5jzp\" (UniqueName: \"kubernetes.io/projected/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-kube-api-access-p5jzp\") pod \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.826723 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-nb\") pod \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.827003 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-svc\") pod \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.827101 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-sb\") pod \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.827239 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-swift-storage-0\") pod \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.827431 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-config\") pod \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\" (UID: \"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f\") " Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.838868 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-kube-api-access-p5jzp" (OuterVolumeSpecName: "kube-api-access-p5jzp") pod "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" (UID: "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f"). InnerVolumeSpecName "kube-api-access-p5jzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.931825 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5jzp\" (UniqueName: \"kubernetes.io/projected/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-kube-api-access-p5jzp\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.932975 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" (UID: "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.945674 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" (UID: "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.946115 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" (UID: "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:06 crc kubenswrapper[4715]: I1203 22:19:06.981424 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" (UID: "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.016522 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-config" (OuterVolumeSpecName: "config") pod "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" (UID: "4cf2d8e3-015c-4abd-9d5a-a38347d6c48f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.040826 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.040869 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.040881 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.040892 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.040903 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.158999 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.438761 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.446320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63","Type":"ContainerDied","Data":"38cc7a0eb158b91e3be83a08e5e66f8c93e42ed4a6cb744e1813db66828afdf8"} Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.447069 4715 scope.go:117] "RemoveContainer" containerID="4178b3536c867700181c222e6dcc90b6a769b5e2fb10972b1829e75b67a91c95" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.450420 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" event={"ID":"4cf2d8e3-015c-4abd-9d5a-a38347d6c48f","Type":"ContainerDied","Data":"03407d6c823302ef91a508113cd5ce4e0020e62236ef70f659542a9701916a7c"} Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.450680 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-fhfjs" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.453450 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerStarted","Data":"72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc"} Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.479600 4715 scope.go:117] "RemoveContainer" containerID="6c1d67be0f901af6d566732377d32d0ce6ce7f3ae85a61efffcf6c009a643293" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.529283 4715 scope.go:117] "RemoveContainer" containerID="d1971a09d49c8130bc0a70f573319660f3e7e1990d43eafef19855a34f28f459" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.529664 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.565905 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.588619 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:19:07 crc kubenswrapper[4715]: E1203 22:19:07.589107 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" containerName="dnsmasq-dns" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.589180 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" containerName="dnsmasq-dns" Dec 03 22:19:07 crc kubenswrapper[4715]: E1203 22:19:07.589243 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" containerName="init" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.589312 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" containerName="init" Dec 03 22:19:07 crc kubenswrapper[4715]: E1203 22:19:07.589383 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerName="nova-metadata-log" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.589437 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerName="nova-metadata-log" Dec 03 22:19:07 crc kubenswrapper[4715]: E1203 22:19:07.589492 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerName="nova-metadata-metadata" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.589571 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerName="nova-metadata-metadata" Dec 03 22:19:07 crc kubenswrapper[4715]: E1203 22:19:07.589632 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ac13b78-4213-452d-8619-9a2bc7ee5338" containerName="nova-manage" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.589685 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ac13b78-4213-452d-8619-9a2bc7ee5338" containerName="nova-manage" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.589921 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ac13b78-4213-452d-8619-9a2bc7ee5338" containerName="nova-manage" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.590002 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerName="nova-metadata-log" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.590066 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" containerName="dnsmasq-dns" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.590119 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" containerName="nova-metadata-metadata" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.591071 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.591210 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.593054 4715 scope.go:117] "RemoveContainer" containerID="bd19da69710577d4336bbfe6bd35bf47ed5e47784c2fffa80d530f10e427e6de" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.596208 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-fhfjs"] Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.602991 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.603188 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.608740 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-fhfjs"] Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.683768 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-config-data\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.683843 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-logs\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.683865 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgl5p\" (UniqueName: \"kubernetes.io/projected/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-kube-api-access-rgl5p\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.683896 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.683941 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.713681 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63" path="/var/lib/kubelet/pods/1b4f3d2f-1d48-46db-b9a5-e5a0b6c7ec63/volumes" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.714464 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cf2d8e3-015c-4abd-9d5a-a38347d6c48f" path="/var/lib/kubelet/pods/4cf2d8e3-015c-4abd-9d5a-a38347d6c48f/volumes" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.787750 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.787892 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-config-data\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.787929 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-logs\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.787948 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgl5p\" (UniqueName: \"kubernetes.io/projected/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-kube-api-access-rgl5p\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.787979 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.789040 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-logs\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.795866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.797123 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-config-data\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.799017 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.823517 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgl5p\" (UniqueName: \"kubernetes.io/projected/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-kube-api-access-rgl5p\") pod \"nova-metadata-0\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " pod="openstack/nova-metadata-0" Dec 03 22:19:07 crc kubenswrapper[4715]: I1203 22:19:07.936668 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:19:08 crc kubenswrapper[4715]: I1203 22:19:08.468235 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:19:08 crc kubenswrapper[4715]: I1203 22:19:08.468762 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-log" containerID="cri-o://2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92" gracePeriod=30 Dec 03 22:19:08 crc kubenswrapper[4715]: I1203 22:19:08.468905 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f199e727-bde9-4b7f-9c09-e5b17a13beda" containerName="nova-scheduler-scheduler" containerID="cri-o://77231f7728a88f0edb7e29432c877cd91f0f46986c5b237eef2a47a80c138f42" gracePeriod=30 Dec 03 22:19:08 crc kubenswrapper[4715]: I1203 22:19:08.469130 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-api" containerID="cri-o://52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05" gracePeriod=30 Dec 03 22:19:09 crc kubenswrapper[4715]: I1203 22:19:09.487434 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2533a7c4-9e84-4f23-b2d5-065ca474fbb4","Type":"ContainerStarted","Data":"8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f"} Dec 03 22:19:09 crc kubenswrapper[4715]: I1203 22:19:09.487887 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2533a7c4-9e84-4f23-b2d5-065ca474fbb4","Type":"ContainerStarted","Data":"810c881ec3f811504782179593158c21ae7f2e54316940ee414680cfb993cf4b"} Dec 03 22:19:09 crc kubenswrapper[4715]: I1203 22:19:09.490863 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerStarted","Data":"c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85"} Dec 03 22:19:09 crc kubenswrapper[4715]: I1203 22:19:09.492705 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 22:19:09 crc kubenswrapper[4715]: I1203 22:19:09.497221 4715 generic.go:334] "Generic (PLEG): container finished" podID="5d7be858-2aa9-45ff-b195-36738f727990" containerID="2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92" exitCode=143 Dec 03 22:19:09 crc kubenswrapper[4715]: I1203 22:19:09.497258 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5d7be858-2aa9-45ff-b195-36738f727990","Type":"ContainerDied","Data":"2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92"} Dec 03 22:19:09 crc kubenswrapper[4715]: I1203 22:19:09.518776 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.928627939 podStartE2EDuration="7.518759171s" podCreationTimestamp="2025-12-03 22:19:02 +0000 UTC" firstStartedPulling="2025-12-03 22:19:03.949241197 +0000 UTC m=+2240.691951812" lastFinishedPulling="2025-12-03 22:19:08.539372449 +0000 UTC m=+2245.282083044" observedRunningTime="2025-12-03 22:19:09.513732086 +0000 UTC m=+2246.256442761" watchObservedRunningTime="2025-12-03 22:19:09.518759171 +0000 UTC m=+2246.261469776" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.511164 4715 generic.go:334] "Generic (PLEG): container finished" podID="f199e727-bde9-4b7f-9c09-e5b17a13beda" containerID="77231f7728a88f0edb7e29432c877cd91f0f46986c5b237eef2a47a80c138f42" exitCode=0 Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.511259 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f199e727-bde9-4b7f-9c09-e5b17a13beda","Type":"ContainerDied","Data":"77231f7728a88f0edb7e29432c877cd91f0f46986c5b237eef2a47a80c138f42"} Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.511534 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f199e727-bde9-4b7f-9c09-e5b17a13beda","Type":"ContainerDied","Data":"993f96d33d12b942b4c672d496922a3be91e23dd59963871dd4c3ac1fbf060d4"} Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.511553 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="993f96d33d12b942b4c672d496922a3be91e23dd59963871dd4c3ac1fbf060d4" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.514471 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2533a7c4-9e84-4f23-b2d5-065ca474fbb4","Type":"ContainerStarted","Data":"d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2"} Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.520102 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.536937 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.536918039 podStartE2EDuration="3.536918039s" podCreationTimestamp="2025-12-03 22:19:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:10.53324538 +0000 UTC m=+2247.275956015" watchObservedRunningTime="2025-12-03 22:19:10.536918039 +0000 UTC m=+2247.279628634" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.648634 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-config-data\") pod \"f199e727-bde9-4b7f-9c09-e5b17a13beda\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.649936 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-combined-ca-bundle\") pod \"f199e727-bde9-4b7f-9c09-e5b17a13beda\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.650008 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ck9hn\" (UniqueName: \"kubernetes.io/projected/f199e727-bde9-4b7f-9c09-e5b17a13beda-kube-api-access-ck9hn\") pod \"f199e727-bde9-4b7f-9c09-e5b17a13beda\" (UID: \"f199e727-bde9-4b7f-9c09-e5b17a13beda\") " Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.661435 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f199e727-bde9-4b7f-9c09-e5b17a13beda-kube-api-access-ck9hn" (OuterVolumeSpecName: "kube-api-access-ck9hn") pod "f199e727-bde9-4b7f-9c09-e5b17a13beda" (UID: "f199e727-bde9-4b7f-9c09-e5b17a13beda"). InnerVolumeSpecName "kube-api-access-ck9hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.687183 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-config-data" (OuterVolumeSpecName: "config-data") pod "f199e727-bde9-4b7f-9c09-e5b17a13beda" (UID: "f199e727-bde9-4b7f-9c09-e5b17a13beda"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.692428 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f199e727-bde9-4b7f-9c09-e5b17a13beda" (UID: "f199e727-bde9-4b7f-9c09-e5b17a13beda"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.752353 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.752399 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ck9hn\" (UniqueName: \"kubernetes.io/projected/f199e727-bde9-4b7f-9c09-e5b17a13beda-kube-api-access-ck9hn\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:10 crc kubenswrapper[4715]: I1203 22:19:10.752412 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f199e727-bde9-4b7f-9c09-e5b17a13beda-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.523752 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.562566 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.580829 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.617090 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:19:11 crc kubenswrapper[4715]: E1203 22:19:11.617756 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f199e727-bde9-4b7f-9c09-e5b17a13beda" containerName="nova-scheduler-scheduler" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.617779 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f199e727-bde9-4b7f-9c09-e5b17a13beda" containerName="nova-scheduler-scheduler" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.618015 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f199e727-bde9-4b7f-9c09-e5b17a13beda" containerName="nova-scheduler-scheduler" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.618881 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.623186 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.653707 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f199e727-bde9-4b7f-9c09-e5b17a13beda" path="/var/lib/kubelet/pods/f199e727-bde9-4b7f-9c09-e5b17a13beda/volumes" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.690490 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n42d5\" (UniqueName: \"kubernetes.io/projected/f7481a5f-d565-4ac9-b629-bfb792cb54f2-kube-api-access-n42d5\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.690602 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.690730 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-config-data\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.690827 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.796433 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-config-data\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.797324 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n42d5\" (UniqueName: \"kubernetes.io/projected/f7481a5f-d565-4ac9-b629-bfb792cb54f2-kube-api-access-n42d5\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.798625 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.802696 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.804598 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-config-data\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:11 crc kubenswrapper[4715]: I1203 22:19:11.817951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n42d5\" (UniqueName: \"kubernetes.io/projected/f7481a5f-d565-4ac9-b629-bfb792cb54f2-kube-api-access-n42d5\") pod \"nova-scheduler-0\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " pod="openstack/nova-scheduler-0" Dec 03 22:19:12 crc kubenswrapper[4715]: I1203 22:19:12.022103 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:19:12 crc kubenswrapper[4715]: I1203 22:19:12.527415 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:19:12 crc kubenswrapper[4715]: I1203 22:19:12.535226 4715 generic.go:334] "Generic (PLEG): container finished" podID="f05149da-e2c3-4bbb-9b7b-e51f6d022875" containerID="d38a7035297765d245716784ed4f49f1e33b03bdd2f0cd344fa42d5fddd3f62a" exitCode=0 Dec 03 22:19:12 crc kubenswrapper[4715]: I1203 22:19:12.535283 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-svdtv" event={"ID":"f05149da-e2c3-4bbb-9b7b-e51f6d022875","Type":"ContainerDied","Data":"d38a7035297765d245716784ed4f49f1e33b03bdd2f0cd344fa42d5fddd3f62a"} Dec 03 22:19:12 crc kubenswrapper[4715]: W1203 22:19:12.544251 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7481a5f_d565_4ac9_b629_bfb792cb54f2.slice/crio-064a1d5770824c6acc1a08fd16fe5aaa6a7be27072b1c498d784b9506ba5e91b WatchSource:0}: Error finding container 064a1d5770824c6acc1a08fd16fe5aaa6a7be27072b1c498d784b9506ba5e91b: Status 404 returned error can't find the container with id 064a1d5770824c6acc1a08fd16fe5aaa6a7be27072b1c498d784b9506ba5e91b Dec 03 22:19:12 crc kubenswrapper[4715]: I1203 22:19:12.937207 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 22:19:12 crc kubenswrapper[4715]: I1203 22:19:12.937864 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.436198 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.546803 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7be858-2aa9-45ff-b195-36738f727990-logs\") pod \"5d7be858-2aa9-45ff-b195-36738f727990\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.546887 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pkk8q\" (UniqueName: \"kubernetes.io/projected/5d7be858-2aa9-45ff-b195-36738f727990-kube-api-access-pkk8q\") pod \"5d7be858-2aa9-45ff-b195-36738f727990\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.546934 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-config-data\") pod \"5d7be858-2aa9-45ff-b195-36738f727990\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.546959 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-combined-ca-bundle\") pod \"5d7be858-2aa9-45ff-b195-36738f727990\" (UID: \"5d7be858-2aa9-45ff-b195-36738f727990\") " Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.547932 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d7be858-2aa9-45ff-b195-36738f727990-logs" (OuterVolumeSpecName: "logs") pod "5d7be858-2aa9-45ff-b195-36738f727990" (UID: "5d7be858-2aa9-45ff-b195-36738f727990"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.548577 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5d7be858-2aa9-45ff-b195-36738f727990-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.555000 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d7be858-2aa9-45ff-b195-36738f727990-kube-api-access-pkk8q" (OuterVolumeSpecName: "kube-api-access-pkk8q") pod "5d7be858-2aa9-45ff-b195-36738f727990" (UID: "5d7be858-2aa9-45ff-b195-36738f727990"). InnerVolumeSpecName "kube-api-access-pkk8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.567422 4715 generic.go:334] "Generic (PLEG): container finished" podID="5d7be858-2aa9-45ff-b195-36738f727990" containerID="52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05" exitCode=0 Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.567816 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.567799 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5d7be858-2aa9-45ff-b195-36738f727990","Type":"ContainerDied","Data":"52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05"} Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.568002 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5d7be858-2aa9-45ff-b195-36738f727990","Type":"ContainerDied","Data":"32feb3e913fa766ae016ef277721d615e27a9171666003db6977f3df2157bfc8"} Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.568050 4715 scope.go:117] "RemoveContainer" containerID="52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.576826 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7481a5f-d565-4ac9-b629-bfb792cb54f2","Type":"ContainerStarted","Data":"c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227"} Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.576895 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7481a5f-d565-4ac9-b629-bfb792cb54f2","Type":"ContainerStarted","Data":"064a1d5770824c6acc1a08fd16fe5aaa6a7be27072b1c498d784b9506ba5e91b"} Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.601834 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d7be858-2aa9-45ff-b195-36738f727990" (UID: "5d7be858-2aa9-45ff-b195-36738f727990"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.601940 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-config-data" (OuterVolumeSpecName: "config-data") pod "5d7be858-2aa9-45ff-b195-36738f727990" (UID: "5d7be858-2aa9-45ff-b195-36738f727990"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.603015 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.603001534 podStartE2EDuration="2.603001534s" podCreationTimestamp="2025-12-03 22:19:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:13.595787761 +0000 UTC m=+2250.338498356" watchObservedRunningTime="2025-12-03 22:19:13.603001534 +0000 UTC m=+2250.345712129" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.618990 4715 scope.go:117] "RemoveContainer" containerID="2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.648536 4715 scope.go:117] "RemoveContainer" containerID="52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05" Dec 03 22:19:13 crc kubenswrapper[4715]: E1203 22:19:13.648999 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05\": container with ID starting with 52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05 not found: ID does not exist" containerID="52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.649084 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05"} err="failed to get container status \"52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05\": rpc error: code = NotFound desc = could not find container \"52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05\": container with ID starting with 52e21191b312d4943cea8c65b26cb3a9e69d0616b284e7f93c62b8e1168efa05 not found: ID does not exist" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.649146 4715 scope.go:117] "RemoveContainer" containerID="2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92" Dec 03 22:19:13 crc kubenswrapper[4715]: E1203 22:19:13.649519 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92\": container with ID starting with 2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92 not found: ID does not exist" containerID="2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.649547 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92"} err="failed to get container status \"2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92\": rpc error: code = NotFound desc = could not find container \"2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92\": container with ID starting with 2b19b1e17ca29f1937396b31628ab05da6f58ba666621278133f7508f02c8a92 not found: ID does not exist" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.650134 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pkk8q\" (UniqueName: \"kubernetes.io/projected/5d7be858-2aa9-45ff-b195-36738f727990-kube-api-access-pkk8q\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.650157 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.650165 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d7be858-2aa9-45ff-b195-36738f727990-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.942621 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:19:13 crc kubenswrapper[4715]: I1203 22:19:13.970218 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.015782 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.056418 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.060893 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqpdd\" (UniqueName: \"kubernetes.io/projected/f05149da-e2c3-4bbb-9b7b-e51f6d022875-kube-api-access-sqpdd\") pod \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063120 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-config-data\") pod \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063215 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-combined-ca-bundle\") pod \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063282 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-scripts\") pod \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\" (UID: \"f05149da-e2c3-4bbb-9b7b-e51f6d022875\") " Dec 03 22:19:14 crc kubenswrapper[4715]: E1203 22:19:14.063296 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-api" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063317 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-api" Dec 03 22:19:14 crc kubenswrapper[4715]: E1203 22:19:14.063347 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f05149da-e2c3-4bbb-9b7b-e51f6d022875" containerName="nova-cell1-conductor-db-sync" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063353 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f05149da-e2c3-4bbb-9b7b-e51f6d022875" containerName="nova-cell1-conductor-db-sync" Dec 03 22:19:14 crc kubenswrapper[4715]: E1203 22:19:14.063378 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-log" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063384 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-log" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063619 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-api" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063634 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d7be858-2aa9-45ff-b195-36738f727990" containerName="nova-api-log" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.063649 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f05149da-e2c3-4bbb-9b7b-e51f6d022875" containerName="nova-cell1-conductor-db-sync" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.067781 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f05149da-e2c3-4bbb-9b7b-e51f6d022875-kube-api-access-sqpdd" (OuterVolumeSpecName: "kube-api-access-sqpdd") pod "f05149da-e2c3-4bbb-9b7b-e51f6d022875" (UID: "f05149da-e2c3-4bbb-9b7b-e51f6d022875"). InnerVolumeSpecName "kube-api-access-sqpdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.068658 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.068746 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.073941 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-scripts" (OuterVolumeSpecName: "scripts") pod "f05149da-e2c3-4bbb-9b7b-e51f6d022875" (UID: "f05149da-e2c3-4bbb-9b7b-e51f6d022875"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.074427 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.098732 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-config-data" (OuterVolumeSpecName: "config-data") pod "f05149da-e2c3-4bbb-9b7b-e51f6d022875" (UID: "f05149da-e2c3-4bbb-9b7b-e51f6d022875"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.099167 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f05149da-e2c3-4bbb-9b7b-e51f6d022875" (UID: "f05149da-e2c3-4bbb-9b7b-e51f6d022875"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.167390 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354af69a-d19f-4d6e-97e4-354ee59d20fc-logs\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.167581 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.167610 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4wc8\" (UniqueName: \"kubernetes.io/projected/354af69a-d19f-4d6e-97e4-354ee59d20fc-kube-api-access-w4wc8\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.167699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-config-data\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.168067 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.168088 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.168099 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f05149da-e2c3-4bbb-9b7b-e51f6d022875-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.168111 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqpdd\" (UniqueName: \"kubernetes.io/projected/f05149da-e2c3-4bbb-9b7b-e51f6d022875-kube-api-access-sqpdd\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.270004 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354af69a-d19f-4d6e-97e4-354ee59d20fc-logs\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.270096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.270117 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4wc8\" (UniqueName: \"kubernetes.io/projected/354af69a-d19f-4d6e-97e4-354ee59d20fc-kube-api-access-w4wc8\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.270137 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-config-data\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.270799 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354af69a-d19f-4d6e-97e4-354ee59d20fc-logs\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.273586 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-config-data\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.274205 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.303692 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4wc8\" (UniqueName: \"kubernetes.io/projected/354af69a-d19f-4d6e-97e4-354ee59d20fc-kube-api-access-w4wc8\") pod \"nova-api-0\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.478069 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.594243 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-svdtv" event={"ID":"f05149da-e2c3-4bbb-9b7b-e51f6d022875","Type":"ContainerDied","Data":"afb6f288619a16001376d7932b860c58dc1a1f8417edc57c3f78d64800905012"} Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.594314 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afb6f288619a16001376d7932b860c58dc1a1f8417edc57c3f78d64800905012" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.594269 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-svdtv" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.686587 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.688196 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.691299 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.696217 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.778233 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b686fe6-ab19-4e75-9328-7ae61b8b609f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.778273 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b686fe6-ab19-4e75-9328-7ae61b8b609f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.778337 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tf7lb\" (UniqueName: \"kubernetes.io/projected/5b686fe6-ab19-4e75-9328-7ae61b8b609f-kube-api-access-tf7lb\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.880350 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tf7lb\" (UniqueName: \"kubernetes.io/projected/5b686fe6-ab19-4e75-9328-7ae61b8b609f-kube-api-access-tf7lb\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.880544 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b686fe6-ab19-4e75-9328-7ae61b8b609f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.880567 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b686fe6-ab19-4e75-9328-7ae61b8b609f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.889539 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b686fe6-ab19-4e75-9328-7ae61b8b609f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.890303 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b686fe6-ab19-4e75-9328-7ae61b8b609f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:14 crc kubenswrapper[4715]: I1203 22:19:14.898437 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tf7lb\" (UniqueName: \"kubernetes.io/projected/5b686fe6-ab19-4e75-9328-7ae61b8b609f-kube-api-access-tf7lb\") pod \"nova-cell1-conductor-0\" (UID: \"5b686fe6-ab19-4e75-9328-7ae61b8b609f\") " pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:15 crc kubenswrapper[4715]: I1203 22:19:15.024113 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:15 crc kubenswrapper[4715]: I1203 22:19:15.066574 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:15 crc kubenswrapper[4715]: E1203 22:19:15.133158 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice/crio-2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice\": RecentStats: unable to find data in memory cache]" Dec 03 22:19:15 crc kubenswrapper[4715]: W1203 22:19:15.502335 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b686fe6_ab19_4e75_9328_7ae61b8b609f.slice/crio-f09d94604a6fdea2c80911b93f8cba40739c39a30102a002bd87c3ab58c5628d WatchSource:0}: Error finding container f09d94604a6fdea2c80911b93f8cba40739c39a30102a002bd87c3ab58c5628d: Status 404 returned error can't find the container with id f09d94604a6fdea2c80911b93f8cba40739c39a30102a002bd87c3ab58c5628d Dec 03 22:19:15 crc kubenswrapper[4715]: I1203 22:19:15.517131 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 22:19:15 crc kubenswrapper[4715]: I1203 22:19:15.632682 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354af69a-d19f-4d6e-97e4-354ee59d20fc","Type":"ContainerStarted","Data":"87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346"} Dec 03 22:19:15 crc kubenswrapper[4715]: I1203 22:19:15.632726 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354af69a-d19f-4d6e-97e4-354ee59d20fc","Type":"ContainerStarted","Data":"d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182"} Dec 03 22:19:15 crc kubenswrapper[4715]: I1203 22:19:15.632737 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354af69a-d19f-4d6e-97e4-354ee59d20fc","Type":"ContainerStarted","Data":"0e58a57b2008cd193881c98d7cae847b8ff115e143c73d322b28e948a6d9cca8"} Dec 03 22:19:15 crc kubenswrapper[4715]: I1203 22:19:15.658746 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d7be858-2aa9-45ff-b195-36738f727990" path="/var/lib/kubelet/pods/5d7be858-2aa9-45ff-b195-36738f727990/volumes" Dec 03 22:19:15 crc kubenswrapper[4715]: I1203 22:19:15.659572 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5b686fe6-ab19-4e75-9328-7ae61b8b609f","Type":"ContainerStarted","Data":"f09d94604a6fdea2c80911b93f8cba40739c39a30102a002bd87c3ab58c5628d"} Dec 03 22:19:16 crc kubenswrapper[4715]: I1203 22:19:16.676967 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"5b686fe6-ab19-4e75-9328-7ae61b8b609f","Type":"ContainerStarted","Data":"9c2c4ab4f45be6a2ef97c21b6360244c9aca55dcd8c8fcca9dc8ccdd97509e48"} Dec 03 22:19:16 crc kubenswrapper[4715]: I1203 22:19:16.677593 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:16 crc kubenswrapper[4715]: I1203 22:19:16.697648 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.697628192 podStartE2EDuration="2.697628192s" podCreationTimestamp="2025-12-03 22:19:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:16.695493924 +0000 UTC m=+2253.438204519" watchObservedRunningTime="2025-12-03 22:19:16.697628192 +0000 UTC m=+2253.440338797" Dec 03 22:19:16 crc kubenswrapper[4715]: I1203 22:19:16.707623 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.707602328 podStartE2EDuration="3.707602328s" podCreationTimestamp="2025-12-03 22:19:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:15.665838899 +0000 UTC m=+2252.408549494" watchObservedRunningTime="2025-12-03 22:19:16.707602328 +0000 UTC m=+2253.450312933" Dec 03 22:19:17 crc kubenswrapper[4715]: I1203 22:19:17.023185 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 22:19:17 crc kubenswrapper[4715]: I1203 22:19:17.937686 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 22:19:17 crc kubenswrapper[4715]: I1203 22:19:17.938192 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 22:19:18 crc kubenswrapper[4715]: I1203 22:19:18.959957 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 22:19:18 crc kubenswrapper[4715]: I1203 22:19:18.960465 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 22:19:19 crc kubenswrapper[4715]: I1203 22:19:19.641294 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:19:19 crc kubenswrapper[4715]: E1203 22:19:19.641925 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:19:20 crc kubenswrapper[4715]: I1203 22:19:20.132893 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 22:19:22 crc kubenswrapper[4715]: I1203 22:19:22.027962 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 22:19:22 crc kubenswrapper[4715]: I1203 22:19:22.077934 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 22:19:22 crc kubenswrapper[4715]: I1203 22:19:22.774718 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 22:19:24 crc kubenswrapper[4715]: I1203 22:19:24.479327 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 22:19:24 crc kubenswrapper[4715]: I1203 22:19:24.479620 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 22:19:25 crc kubenswrapper[4715]: E1203 22:19:25.366336 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice/crio-2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice\": RecentStats: unable to find data in memory cache]" Dec 03 22:19:25 crc kubenswrapper[4715]: I1203 22:19:25.561693 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 22:19:25 crc kubenswrapper[4715]: I1203 22:19:25.561721 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.192:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 22:19:27 crc kubenswrapper[4715]: I1203 22:19:27.943040 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 22:19:27 crc kubenswrapper[4715]: I1203 22:19:27.947499 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 22:19:27 crc kubenswrapper[4715]: I1203 22:19:27.949683 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 22:19:28 crc kubenswrapper[4715]: I1203 22:19:28.816710 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 22:19:32 crc kubenswrapper[4715]: I1203 22:19:32.733052 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 22:19:33 crc kubenswrapper[4715]: I1203 22:19:33.643848 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:19:33 crc kubenswrapper[4715]: E1203 22:19:33.644281 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.482691 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.483459 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.483867 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.484180 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.486423 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.488452 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.674737 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-fnbg8"] Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.682043 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.693425 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-fnbg8"] Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.740491 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.743689 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-config\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.743737 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.743792 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dbzm\" (UniqueName: \"kubernetes.io/projected/08ba21f3-cfda-482a-b582-0cf75e86acea-kube-api-access-9dbzm\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.743859 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.743904 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.743932 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.844654 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-config-data\") pod \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.844716 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdpdv\" (UniqueName: \"kubernetes.io/projected/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-kube-api-access-hdpdv\") pod \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.844824 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-combined-ca-bundle\") pod \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\" (UID: \"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8\") " Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.845218 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-config\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.845421 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.845479 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dbzm\" (UniqueName: \"kubernetes.io/projected/08ba21f3-cfda-482a-b582-0cf75e86acea-kube-api-access-9dbzm\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.845541 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.845587 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.845613 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.846949 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-config\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.846969 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.847174 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.848213 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.853061 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.855288 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-kube-api-access-hdpdv" (OuterVolumeSpecName: "kube-api-access-hdpdv") pod "78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" (UID: "78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8"). InnerVolumeSpecName "kube-api-access-hdpdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.862693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dbzm\" (UniqueName: \"kubernetes.io/projected/08ba21f3-cfda-482a-b582-0cf75e86acea-kube-api-access-9dbzm\") pod \"dnsmasq-dns-cd5cbd7b9-fnbg8\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.873846 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-config-data" (OuterVolumeSpecName: "config-data") pod "78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" (UID: "78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.874954 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" (UID: "78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.887834 4715 generic.go:334] "Generic (PLEG): container finished" podID="78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" containerID="57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d" exitCode=137 Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.888662 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.896751 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8","Type":"ContainerDied","Data":"57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d"} Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.896782 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8","Type":"ContainerDied","Data":"4a7aadbfd0f4303461a5231d71ec42aadf85a4ccc558ae0480b4dc38cbfb4dc7"} Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.896799 4715 scope.go:117] "RemoveContainer" containerID="57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.947627 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.947655 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.947664 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdpdv\" (UniqueName: \"kubernetes.io/projected/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8-kube-api-access-hdpdv\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.950647 4715 scope.go:117] "RemoveContainer" containerID="57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d" Dec 03 22:19:34 crc kubenswrapper[4715]: E1203 22:19:34.951223 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d\": container with ID starting with 57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d not found: ID does not exist" containerID="57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.951260 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d"} err="failed to get container status \"57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d\": rpc error: code = NotFound desc = could not find container \"57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d\": container with ID starting with 57319c4f9a52dbec5232a86758c9bba6fed175ab0f25097e7a7d730956e1415d not found: ID does not exist" Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.959548 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:19:34 crc kubenswrapper[4715]: I1203 22:19:34.988742 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.021639 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:19:35 crc kubenswrapper[4715]: E1203 22:19:35.022068 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.022085 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.022259 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.022902 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.025452 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.028302 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.028479 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.031579 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.048061 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.048120 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.048438 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-md7gn\" (UniqueName: \"kubernetes.io/projected/e4f46a99-13e6-4175-939a-3ffb54503ef2-kube-api-access-md7gn\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.048513 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.048590 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.055122 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.150244 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.150337 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-md7gn\" (UniqueName: \"kubernetes.io/projected/e4f46a99-13e6-4175-939a-3ffb54503ef2-kube-api-access-md7gn\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.150358 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.150386 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.150428 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.156402 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.156429 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.157184 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.180122 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4f46a99-13e6-4175-939a-3ffb54503ef2-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.185179 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-md7gn\" (UniqueName: \"kubernetes.io/projected/e4f46a99-13e6-4175-939a-3ffb54503ef2-kube-api-access-md7gn\") pod \"nova-cell1-novncproxy-0\" (UID: \"e4f46a99-13e6-4175-939a-3ffb54503ef2\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.340848 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.488757 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-fnbg8"] Dec 03 22:19:35 crc kubenswrapper[4715]: E1203 22:19:35.621798 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod008014fe_5f97_4de4_a7ef_4bf119aed38f.slice/crio-2a5d1b1197c1124c49b25d473c04364c0c6ba632a4d9bc0b3e431c061239f71f\": RecentStats: unable to find data in memory cache]" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.666992 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8" path="/var/lib/kubelet/pods/78f4d687-a0a9-49ef-86a6-4e9ebf6af0f8/volumes" Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.806592 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.900797 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e4f46a99-13e6-4175-939a-3ffb54503ef2","Type":"ContainerStarted","Data":"a5fe07845a9fdea131b065eca6291e258bdc7610b9c8d28c14196b2d22eaeec9"} Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.904184 4715 generic.go:334] "Generic (PLEG): container finished" podID="08ba21f3-cfda-482a-b582-0cf75e86acea" containerID="94d628406d5f2b5b4249b4c5cff8bbb9a42ec6c461ad86713cf2a8412bcf2a89" exitCode=0 Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.904294 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" event={"ID":"08ba21f3-cfda-482a-b582-0cf75e86acea","Type":"ContainerDied","Data":"94d628406d5f2b5b4249b4c5cff8bbb9a42ec6c461ad86713cf2a8412bcf2a89"} Dec 03 22:19:35 crc kubenswrapper[4715]: I1203 22:19:35.904362 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" event={"ID":"08ba21f3-cfda-482a-b582-0cf75e86acea","Type":"ContainerStarted","Data":"16de09fd2273181d56a5983eb4929dbb0f1ab5dbef3ac925d77cc4f449653c4a"} Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.684400 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.916773 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" event={"ID":"08ba21f3-cfda-482a-b582-0cf75e86acea","Type":"ContainerStarted","Data":"2d6390320ff45b454468fb873b15d44f83cd163844da43b681ba5423756ce064"} Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.917182 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.919835 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e4f46a99-13e6-4175-939a-3ffb54503ef2","Type":"ContainerStarted","Data":"3502f00ec12be3242c27cd120e117fcb4e968d11994fbf69928f0841d36d7624"} Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.919868 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-log" containerID="cri-o://d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182" gracePeriod=30 Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.919964 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-api" containerID="cri-o://87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346" gracePeriod=30 Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.947203 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" podStartSLOduration=2.9471835459999998 podStartE2EDuration="2.947183546s" podCreationTimestamp="2025-12-03 22:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:36.939035239 +0000 UTC m=+2273.681745834" watchObservedRunningTime="2025-12-03 22:19:36.947183546 +0000 UTC m=+2273.689894141" Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.959896 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.960176 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="ceilometer-central-agent" containerID="cri-o://5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb" gracePeriod=30 Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.960216 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="sg-core" containerID="cri-o://72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc" gracePeriod=30 Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.960320 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="proxy-httpd" containerID="cri-o://c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85" gracePeriod=30 Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.960565 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="ceilometer-notification-agent" containerID="cri-o://618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef" gracePeriod=30 Dec 03 22:19:36 crc kubenswrapper[4715]: I1203 22:19:36.983132 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.983113797 podStartE2EDuration="2.983113797s" podCreationTimestamp="2025-12-03 22:19:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:36.974838846 +0000 UTC m=+2273.717549441" watchObservedRunningTime="2025-12-03 22:19:36.983113797 +0000 UTC m=+2273.725824392" Dec 03 22:19:37 crc kubenswrapper[4715]: I1203 22:19:37.960207 4715 generic.go:334] "Generic (PLEG): container finished" podID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerID="d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182" exitCode=143 Dec 03 22:19:37 crc kubenswrapper[4715]: I1203 22:19:37.961055 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354af69a-d19f-4d6e-97e4-354ee59d20fc","Type":"ContainerDied","Data":"d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182"} Dec 03 22:19:37 crc kubenswrapper[4715]: I1203 22:19:37.966039 4715 generic.go:334] "Generic (PLEG): container finished" podID="f582783f-a94f-4910-962b-51f24976b3e6" containerID="c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85" exitCode=0 Dec 03 22:19:37 crc kubenswrapper[4715]: I1203 22:19:37.966061 4715 generic.go:334] "Generic (PLEG): container finished" podID="f582783f-a94f-4910-962b-51f24976b3e6" containerID="72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc" exitCode=2 Dec 03 22:19:37 crc kubenswrapper[4715]: I1203 22:19:37.966068 4715 generic.go:334] "Generic (PLEG): container finished" podID="f582783f-a94f-4910-962b-51f24976b3e6" containerID="5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb" exitCode=0 Dec 03 22:19:37 crc kubenswrapper[4715]: I1203 22:19:37.966125 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerDied","Data":"c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85"} Dec 03 22:19:37 crc kubenswrapper[4715]: I1203 22:19:37.966157 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerDied","Data":"72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc"} Dec 03 22:19:37 crc kubenswrapper[4715]: I1203 22:19:37.966167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerDied","Data":"5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb"} Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.646407 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.838636 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-scripts\") pod \"f582783f-a94f-4910-962b-51f24976b3e6\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.838784 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-log-httpd\") pod \"f582783f-a94f-4910-962b-51f24976b3e6\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.838828 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-config-data\") pod \"f582783f-a94f-4910-962b-51f24976b3e6\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.838880 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-ceilometer-tls-certs\") pod \"f582783f-a94f-4910-962b-51f24976b3e6\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.838902 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-sg-core-conf-yaml\") pod \"f582783f-a94f-4910-962b-51f24976b3e6\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.838933 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-combined-ca-bundle\") pod \"f582783f-a94f-4910-962b-51f24976b3e6\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.839043 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-run-httpd\") pod \"f582783f-a94f-4910-962b-51f24976b3e6\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.839110 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wssbf\" (UniqueName: \"kubernetes.io/projected/f582783f-a94f-4910-962b-51f24976b3e6-kube-api-access-wssbf\") pod \"f582783f-a94f-4910-962b-51f24976b3e6\" (UID: \"f582783f-a94f-4910-962b-51f24976b3e6\") " Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.839528 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f582783f-a94f-4910-962b-51f24976b3e6" (UID: "f582783f-a94f-4910-962b-51f24976b3e6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.839648 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f582783f-a94f-4910-962b-51f24976b3e6" (UID: "f582783f-a94f-4910-962b-51f24976b3e6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.839824 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.845774 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f582783f-a94f-4910-962b-51f24976b3e6-kube-api-access-wssbf" (OuterVolumeSpecName: "kube-api-access-wssbf") pod "f582783f-a94f-4910-962b-51f24976b3e6" (UID: "f582783f-a94f-4910-962b-51f24976b3e6"). InnerVolumeSpecName "kube-api-access-wssbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.846292 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-scripts" (OuterVolumeSpecName: "scripts") pod "f582783f-a94f-4910-962b-51f24976b3e6" (UID: "f582783f-a94f-4910-962b-51f24976b3e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.868717 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f582783f-a94f-4910-962b-51f24976b3e6" (UID: "f582783f-a94f-4910-962b-51f24976b3e6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.914024 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f582783f-a94f-4910-962b-51f24976b3e6" (UID: "f582783f-a94f-4910-962b-51f24976b3e6"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.941989 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f582783f-a94f-4910-962b-51f24976b3e6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.942044 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.942060 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.942073 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wssbf\" (UniqueName: \"kubernetes.io/projected/f582783f-a94f-4910-962b-51f24976b3e6-kube-api-access-wssbf\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.942085 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.948876 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f582783f-a94f-4910-962b-51f24976b3e6" (UID: "f582783f-a94f-4910-962b-51f24976b3e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.967790 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-config-data" (OuterVolumeSpecName: "config-data") pod "f582783f-a94f-4910-962b-51f24976b3e6" (UID: "f582783f-a94f-4910-962b-51f24976b3e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.984726 4715 generic.go:334] "Generic (PLEG): container finished" podID="f582783f-a94f-4910-962b-51f24976b3e6" containerID="618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef" exitCode=0 Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.984770 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerDied","Data":"618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef"} Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.984797 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f582783f-a94f-4910-962b-51f24976b3e6","Type":"ContainerDied","Data":"f73f0918d87642ae79cd9cbc9f4f3ea58d09f028f416d3995a4d531825df670f"} Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.984816 4715 scope.go:117] "RemoveContainer" containerID="c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85" Dec 03 22:19:38 crc kubenswrapper[4715]: I1203 22:19:38.984942 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.024589 4715 scope.go:117] "RemoveContainer" containerID="72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.031801 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.044524 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.044555 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f582783f-a94f-4910-962b-51f24976b3e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.050017 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.063203 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:39 crc kubenswrapper[4715]: E1203 22:19:39.064419 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="ceilometer-central-agent" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064442 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="ceilometer-central-agent" Dec 03 22:19:39 crc kubenswrapper[4715]: E1203 22:19:39.064464 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="ceilometer-notification-agent" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064472 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="ceilometer-notification-agent" Dec 03 22:19:39 crc kubenswrapper[4715]: E1203 22:19:39.064497 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="proxy-httpd" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064516 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="proxy-httpd" Dec 03 22:19:39 crc kubenswrapper[4715]: E1203 22:19:39.064529 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="sg-core" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064535 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="sg-core" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064630 4715 scope.go:117] "RemoveContainer" containerID="618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064933 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="proxy-httpd" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064957 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="sg-core" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064966 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="ceilometer-central-agent" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.064980 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f582783f-a94f-4910-962b-51f24976b3e6" containerName="ceilometer-notification-agent" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.071270 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.074056 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.074276 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.074451 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.077527 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.096209 4715 scope.go:117] "RemoveContainer" containerID="5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.118987 4715 scope.go:117] "RemoveContainer" containerID="c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85" Dec 03 22:19:39 crc kubenswrapper[4715]: E1203 22:19:39.119623 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85\": container with ID starting with c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85 not found: ID does not exist" containerID="c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.119673 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85"} err="failed to get container status \"c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85\": rpc error: code = NotFound desc = could not find container \"c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85\": container with ID starting with c1175db0955e8811059176d5f304002016241ecdbeed4bfdcd155861a5918b85 not found: ID does not exist" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.119711 4715 scope.go:117] "RemoveContainer" containerID="72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc" Dec 03 22:19:39 crc kubenswrapper[4715]: E1203 22:19:39.120173 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc\": container with ID starting with 72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc not found: ID does not exist" containerID="72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.120219 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc"} err="failed to get container status \"72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc\": rpc error: code = NotFound desc = could not find container \"72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc\": container with ID starting with 72f50c27981de9ab6fb6afa23fe9a95691bc0cdf08672e6ac00e42c67a1045bc not found: ID does not exist" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.120249 4715 scope.go:117] "RemoveContainer" containerID="618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef" Dec 03 22:19:39 crc kubenswrapper[4715]: E1203 22:19:39.120609 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef\": container with ID starting with 618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef not found: ID does not exist" containerID="618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.120640 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef"} err="failed to get container status \"618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef\": rpc error: code = NotFound desc = could not find container \"618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef\": container with ID starting with 618ef331a8f90c7f5cc1b8f667f6894c4348824965a6f15d11ea1ef4f2ba9fef not found: ID does not exist" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.120658 4715 scope.go:117] "RemoveContainer" containerID="5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb" Dec 03 22:19:39 crc kubenswrapper[4715]: E1203 22:19:39.120956 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb\": container with ID starting with 5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb not found: ID does not exist" containerID="5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.120987 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb"} err="failed to get container status \"5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb\": rpc error: code = NotFound desc = could not find container \"5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb\": container with ID starting with 5288e41f147e405e7f55b7aa54f376e52354dee5e32af8c950e6918c4e894ccb not found: ID does not exist" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.146153 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpxr4\" (UniqueName: \"kubernetes.io/projected/b35ac264-213f-4e1e-926d-cb6e2b47c684-kube-api-access-dpxr4\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.146319 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.146472 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.146579 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.146675 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-config-data\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.146702 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-log-httpd\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.146842 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-scripts\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.146900 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-run-httpd\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.248545 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-scripts\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.248608 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-run-httpd\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.248687 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpxr4\" (UniqueName: \"kubernetes.io/projected/b35ac264-213f-4e1e-926d-cb6e2b47c684-kube-api-access-dpxr4\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.248760 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.248797 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.248835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.248875 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-config-data\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.248904 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-log-httpd\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.249427 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-run-httpd\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.249924 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-log-httpd\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.252324 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-scripts\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.252605 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.253119 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.254828 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-config-data\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.255008 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.269421 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpxr4\" (UniqueName: \"kubernetes.io/projected/b35ac264-213f-4e1e-926d-cb6e2b47c684-kube-api-access-dpxr4\") pod \"ceilometer-0\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.314535 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.315367 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.658236 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f582783f-a94f-4910-962b-51f24976b3e6" path="/var/lib/kubelet/pods/f582783f-a94f-4910-962b-51f24976b3e6/volumes" Dec 03 22:19:39 crc kubenswrapper[4715]: I1203 22:19:39.805942 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:39 crc kubenswrapper[4715]: W1203 22:19:39.810488 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb35ac264_213f_4e1e_926d_cb6e2b47c684.slice/crio-dfb7e742e29f6074dd1fa35623c2e33f751b24d058067f36b73ed433d08b675e WatchSource:0}: Error finding container dfb7e742e29f6074dd1fa35623c2e33f751b24d058067f36b73ed433d08b675e: Status 404 returned error can't find the container with id dfb7e742e29f6074dd1fa35623c2e33f751b24d058067f36b73ed433d08b675e Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.000881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerStarted","Data":"dfb7e742e29f6074dd1fa35623c2e33f751b24d058067f36b73ed433d08b675e"} Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.342171 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.480327 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.673363 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354af69a-d19f-4d6e-97e4-354ee59d20fc-logs\") pod \"354af69a-d19f-4d6e-97e4-354ee59d20fc\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.673413 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-combined-ca-bundle\") pod \"354af69a-d19f-4d6e-97e4-354ee59d20fc\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.673924 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/354af69a-d19f-4d6e-97e4-354ee59d20fc-logs" (OuterVolumeSpecName: "logs") pod "354af69a-d19f-4d6e-97e4-354ee59d20fc" (UID: "354af69a-d19f-4d6e-97e4-354ee59d20fc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.673985 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4wc8\" (UniqueName: \"kubernetes.io/projected/354af69a-d19f-4d6e-97e4-354ee59d20fc-kube-api-access-w4wc8\") pod \"354af69a-d19f-4d6e-97e4-354ee59d20fc\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.674063 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-config-data\") pod \"354af69a-d19f-4d6e-97e4-354ee59d20fc\" (UID: \"354af69a-d19f-4d6e-97e4-354ee59d20fc\") " Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.674376 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/354af69a-d19f-4d6e-97e4-354ee59d20fc-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.720052 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/354af69a-d19f-4d6e-97e4-354ee59d20fc-kube-api-access-w4wc8" (OuterVolumeSpecName: "kube-api-access-w4wc8") pod "354af69a-d19f-4d6e-97e4-354ee59d20fc" (UID: "354af69a-d19f-4d6e-97e4-354ee59d20fc"). InnerVolumeSpecName "kube-api-access-w4wc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.723614 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-config-data" (OuterVolumeSpecName: "config-data") pod "354af69a-d19f-4d6e-97e4-354ee59d20fc" (UID: "354af69a-d19f-4d6e-97e4-354ee59d20fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.724152 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "354af69a-d19f-4d6e-97e4-354ee59d20fc" (UID: "354af69a-d19f-4d6e-97e4-354ee59d20fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.776192 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.776460 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4wc8\" (UniqueName: \"kubernetes.io/projected/354af69a-d19f-4d6e-97e4-354ee59d20fc-kube-api-access-w4wc8\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:40 crc kubenswrapper[4715]: I1203 22:19:40.776665 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/354af69a-d19f-4d6e-97e4-354ee59d20fc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.010941 4715 generic.go:334] "Generic (PLEG): container finished" podID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerID="87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346" exitCode=0 Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.010996 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.012050 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354af69a-d19f-4d6e-97e4-354ee59d20fc","Type":"ContainerDied","Data":"87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346"} Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.012238 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"354af69a-d19f-4d6e-97e4-354ee59d20fc","Type":"ContainerDied","Data":"0e58a57b2008cd193881c98d7cae847b8ff115e143c73d322b28e948a6d9cca8"} Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.012272 4715 scope.go:117] "RemoveContainer" containerID="87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.013058 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerStarted","Data":"b8403e182d586651bdeb70a45cf83dbd31d91f319e6682852d42f4e8610f95fa"} Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.051912 4715 scope.go:117] "RemoveContainer" containerID="d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.063742 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.078970 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.081333 4715 scope.go:117] "RemoveContainer" containerID="87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346" Dec 03 22:19:41 crc kubenswrapper[4715]: E1203 22:19:41.081767 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346\": container with ID starting with 87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346 not found: ID does not exist" containerID="87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.081799 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346"} err="failed to get container status \"87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346\": rpc error: code = NotFound desc = could not find container \"87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346\": container with ID starting with 87a3e2ae2fc0cf71e448d50caa0624dbbf419b3fb2437aed7096c6e629a3e346 not found: ID does not exist" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.081823 4715 scope.go:117] "RemoveContainer" containerID="d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182" Dec 03 22:19:41 crc kubenswrapper[4715]: E1203 22:19:41.082140 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182\": container with ID starting with d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182 not found: ID does not exist" containerID="d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.082187 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182"} err="failed to get container status \"d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182\": rpc error: code = NotFound desc = could not find container \"d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182\": container with ID starting with d7e251faf124f3c5a0ca57c579fc51ca5e689c6f33fed9768c7a90fbd8bd9182 not found: ID does not exist" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.089090 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:41 crc kubenswrapper[4715]: E1203 22:19:41.089910 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-log" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.090019 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-log" Dec 03 22:19:41 crc kubenswrapper[4715]: E1203 22:19:41.090139 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-api" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.090222 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-api" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.090560 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-log" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.090666 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" containerName="nova-api-api" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.092052 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.096837 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.098025 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.098192 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.123578 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.286264 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.287196 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw6hs\" (UniqueName: \"kubernetes.io/projected/45e6fe90-4116-4017-bd10-3f3b42e68c3e-kube-api-access-bw6hs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.287319 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e6fe90-4116-4017-bd10-3f3b42e68c3e-logs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.287419 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-public-tls-certs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.287521 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-config-data\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.287619 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.389253 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.389717 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw6hs\" (UniqueName: \"kubernetes.io/projected/45e6fe90-4116-4017-bd10-3f3b42e68c3e-kube-api-access-bw6hs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.389762 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e6fe90-4116-4017-bd10-3f3b42e68c3e-logs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.389835 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-public-tls-certs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.389862 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-config-data\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.389891 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.390781 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e6fe90-4116-4017-bd10-3f3b42e68c3e-logs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.393542 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-public-tls-certs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.394326 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.395159 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-config-data\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.395678 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.413826 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw6hs\" (UniqueName: \"kubernetes.io/projected/45e6fe90-4116-4017-bd10-3f3b42e68c3e-kube-api-access-bw6hs\") pod \"nova-api-0\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.419119 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:19:41 crc kubenswrapper[4715]: I1203 22:19:41.649542 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="354af69a-d19f-4d6e-97e4-354ee59d20fc" path="/var/lib/kubelet/pods/354af69a-d19f-4d6e-97e4-354ee59d20fc/volumes" Dec 03 22:19:42 crc kubenswrapper[4715]: I1203 22:19:42.026617 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerStarted","Data":"dab8336003aa53b24b0cf172c6b7088b96dc670fb9b74a66819a0c0367c188f4"} Dec 03 22:19:42 crc kubenswrapper[4715]: I1203 22:19:42.108151 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:42 crc kubenswrapper[4715]: W1203 22:19:42.164223 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45e6fe90_4116_4017_bd10_3f3b42e68c3e.slice/crio-9c77f7dfc17a7c2268e08385dcd03f7bdaba27440acefc9c741b856c2460f710 WatchSource:0}: Error finding container 9c77f7dfc17a7c2268e08385dcd03f7bdaba27440acefc9c741b856c2460f710: Status 404 returned error can't find the container with id 9c77f7dfc17a7c2268e08385dcd03f7bdaba27440acefc9c741b856c2460f710 Dec 03 22:19:43 crc kubenswrapper[4715]: I1203 22:19:43.039571 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45e6fe90-4116-4017-bd10-3f3b42e68c3e","Type":"ContainerStarted","Data":"8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2"} Dec 03 22:19:43 crc kubenswrapper[4715]: I1203 22:19:43.040426 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45e6fe90-4116-4017-bd10-3f3b42e68c3e","Type":"ContainerStarted","Data":"054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86"} Dec 03 22:19:43 crc kubenswrapper[4715]: I1203 22:19:43.040552 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45e6fe90-4116-4017-bd10-3f3b42e68c3e","Type":"ContainerStarted","Data":"9c77f7dfc17a7c2268e08385dcd03f7bdaba27440acefc9c741b856c2460f710"} Dec 03 22:19:43 crc kubenswrapper[4715]: I1203 22:19:43.041117 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerStarted","Data":"86b35addd818249ba60715783d48f08c447ef768a6c26e5d6bf03bfa876b5479"} Dec 03 22:19:43 crc kubenswrapper[4715]: I1203 22:19:43.068725 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.068694342 podStartE2EDuration="2.068694342s" podCreationTimestamp="2025-12-03 22:19:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:43.059186477 +0000 UTC m=+2279.801897082" watchObservedRunningTime="2025-12-03 22:19:43.068694342 +0000 UTC m=+2279.811404957" Dec 03 22:19:44 crc kubenswrapper[4715]: I1203 22:19:44.053141 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="ceilometer-central-agent" containerID="cri-o://b8403e182d586651bdeb70a45cf83dbd31d91f319e6682852d42f4e8610f95fa" gracePeriod=30 Dec 03 22:19:44 crc kubenswrapper[4715]: I1203 22:19:44.053606 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerStarted","Data":"1abc61a4c85ca4fbc980e3269d44a353e556e8a42133803c495d40fa5bc6a9b2"} Dec 03 22:19:44 crc kubenswrapper[4715]: I1203 22:19:44.053737 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 22:19:44 crc kubenswrapper[4715]: I1203 22:19:44.053776 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="proxy-httpd" containerID="cri-o://1abc61a4c85ca4fbc980e3269d44a353e556e8a42133803c495d40fa5bc6a9b2" gracePeriod=30 Dec 03 22:19:44 crc kubenswrapper[4715]: I1203 22:19:44.053944 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="sg-core" containerID="cri-o://86b35addd818249ba60715783d48f08c447ef768a6c26e5d6bf03bfa876b5479" gracePeriod=30 Dec 03 22:19:44 crc kubenswrapper[4715]: I1203 22:19:44.054019 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="ceilometer-notification-agent" containerID="cri-o://dab8336003aa53b24b0cf172c6b7088b96dc670fb9b74a66819a0c0367c188f4" gracePeriod=30 Dec 03 22:19:44 crc kubenswrapper[4715]: I1203 22:19:44.083147 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.6208365 podStartE2EDuration="5.08312787s" podCreationTimestamp="2025-12-03 22:19:39 +0000 UTC" firstStartedPulling="2025-12-03 22:19:39.813277853 +0000 UTC m=+2276.555988438" lastFinishedPulling="2025-12-03 22:19:43.275569183 +0000 UTC m=+2280.018279808" observedRunningTime="2025-12-03 22:19:44.075605719 +0000 UTC m=+2280.818316324" watchObservedRunningTime="2025-12-03 22:19:44.08312787 +0000 UTC m=+2280.825838465" Dec 03 22:19:44 crc kubenswrapper[4715]: I1203 22:19:44.634609 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:19:44 crc kubenswrapper[4715]: E1203 22:19:44.634920 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.057617 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.071940 4715 generic.go:334] "Generic (PLEG): container finished" podID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerID="1abc61a4c85ca4fbc980e3269d44a353e556e8a42133803c495d40fa5bc6a9b2" exitCode=0 Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.071985 4715 generic.go:334] "Generic (PLEG): container finished" podID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerID="86b35addd818249ba60715783d48f08c447ef768a6c26e5d6bf03bfa876b5479" exitCode=2 Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.071996 4715 generic.go:334] "Generic (PLEG): container finished" podID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerID="dab8336003aa53b24b0cf172c6b7088b96dc670fb9b74a66819a0c0367c188f4" exitCode=0 Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.072015 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerDied","Data":"1abc61a4c85ca4fbc980e3269d44a353e556e8a42133803c495d40fa5bc6a9b2"} Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.072038 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerDied","Data":"86b35addd818249ba60715783d48f08c447ef768a6c26e5d6bf03bfa876b5479"} Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.072052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerDied","Data":"dab8336003aa53b24b0cf172c6b7088b96dc670fb9b74a66819a0c0367c188f4"} Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.132436 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-6b88c"] Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.133052 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" podUID="b3f721fb-a909-4614-8f12-ae64db2c78c9" containerName="dnsmasq-dns" containerID="cri-o://6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd" gracePeriod=10 Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.341812 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.368966 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.684547 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.790112 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-nb\") pod \"b3f721fb-a909-4614-8f12-ae64db2c78c9\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.790160 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbgt7\" (UniqueName: \"kubernetes.io/projected/b3f721fb-a909-4614-8f12-ae64db2c78c9-kube-api-access-vbgt7\") pod \"b3f721fb-a909-4614-8f12-ae64db2c78c9\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.790406 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-config\") pod \"b3f721fb-a909-4614-8f12-ae64db2c78c9\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.790443 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-sb\") pod \"b3f721fb-a909-4614-8f12-ae64db2c78c9\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.790463 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-swift-storage-0\") pod \"b3f721fb-a909-4614-8f12-ae64db2c78c9\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.790496 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-svc\") pod \"b3f721fb-a909-4614-8f12-ae64db2c78c9\" (UID: \"b3f721fb-a909-4614-8f12-ae64db2c78c9\") " Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.820811 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3f721fb-a909-4614-8f12-ae64db2c78c9-kube-api-access-vbgt7" (OuterVolumeSpecName: "kube-api-access-vbgt7") pod "b3f721fb-a909-4614-8f12-ae64db2c78c9" (UID: "b3f721fb-a909-4614-8f12-ae64db2c78c9"). InnerVolumeSpecName "kube-api-access-vbgt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.856730 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3f721fb-a909-4614-8f12-ae64db2c78c9" (UID: "b3f721fb-a909-4614-8f12-ae64db2c78c9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.864269 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3f721fb-a909-4614-8f12-ae64db2c78c9" (UID: "b3f721fb-a909-4614-8f12-ae64db2c78c9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.882275 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b3f721fb-a909-4614-8f12-ae64db2c78c9" (UID: "b3f721fb-a909-4614-8f12-ae64db2c78c9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.885096 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3f721fb-a909-4614-8f12-ae64db2c78c9" (UID: "b3f721fb-a909-4614-8f12-ae64db2c78c9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.886269 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-config" (OuterVolumeSpecName: "config") pod "b3f721fb-a909-4614-8f12-ae64db2c78c9" (UID: "b3f721fb-a909-4614-8f12-ae64db2c78c9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.893145 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.893169 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.893180 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.893189 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.893197 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3f721fb-a909-4614-8f12-ae64db2c78c9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.893206 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbgt7\" (UniqueName: \"kubernetes.io/projected/b3f721fb-a909-4614-8f12-ae64db2c78c9-kube-api-access-vbgt7\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.947935 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bwfdk"] Dec 03 22:19:45 crc kubenswrapper[4715]: E1203 22:19:45.948288 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3f721fb-a909-4614-8f12-ae64db2c78c9" containerName="dnsmasq-dns" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.948304 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3f721fb-a909-4614-8f12-ae64db2c78c9" containerName="dnsmasq-dns" Dec 03 22:19:45 crc kubenswrapper[4715]: E1203 22:19:45.948337 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3f721fb-a909-4614-8f12-ae64db2c78c9" containerName="init" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.948346 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3f721fb-a909-4614-8f12-ae64db2c78c9" containerName="init" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.948547 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3f721fb-a909-4614-8f12-ae64db2c78c9" containerName="dnsmasq-dns" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.949787 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:45 crc kubenswrapper[4715]: I1203 22:19:45.964831 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwfdk"] Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.083195 4715 generic.go:334] "Generic (PLEG): container finished" podID="b3f721fb-a909-4614-8f12-ae64db2c78c9" containerID="6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd" exitCode=0 Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.083261 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" event={"ID":"b3f721fb-a909-4614-8f12-ae64db2c78c9","Type":"ContainerDied","Data":"6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd"} Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.083303 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.083316 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-6b88c" event={"ID":"b3f721fb-a909-4614-8f12-ae64db2c78c9","Type":"ContainerDied","Data":"785d959dd2c30c9c023a1d4d3ec32e64c20668144df0af2bbe636223d8019d61"} Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.083334 4715 scope.go:117] "RemoveContainer" containerID="6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.096281 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-catalog-content\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.096339 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpzxz\" (UniqueName: \"kubernetes.io/projected/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-kube-api-access-gpzxz\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.096556 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-utilities\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.111989 4715 scope.go:117] "RemoveContainer" containerID="f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.113068 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.121001 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-6b88c"] Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.129894 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-6b88c"] Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.143011 4715 scope.go:117] "RemoveContainer" containerID="6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd" Dec 03 22:19:46 crc kubenswrapper[4715]: E1203 22:19:46.143781 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd\": container with ID starting with 6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd not found: ID does not exist" containerID="6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.143814 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd"} err="failed to get container status \"6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd\": rpc error: code = NotFound desc = could not find container \"6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd\": container with ID starting with 6d312ba6107ce1dc6d6ff6b43ff12e0eeda450a111d9a4504e3745ff97d335bd not found: ID does not exist" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.143836 4715 scope.go:117] "RemoveContainer" containerID="f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d" Dec 03 22:19:46 crc kubenswrapper[4715]: E1203 22:19:46.144057 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d\": container with ID starting with f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d not found: ID does not exist" containerID="f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.144082 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d"} err="failed to get container status \"f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d\": rpc error: code = NotFound desc = could not find container \"f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d\": container with ID starting with f4d805a55951601ee1adc6da050dec03037ea5f91480bdc3f5a0d8a14820126d not found: ID does not exist" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.210582 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-utilities\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.210764 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-catalog-content\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.210845 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpzxz\" (UniqueName: \"kubernetes.io/projected/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-kube-api-access-gpzxz\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.211783 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-utilities\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.212085 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-catalog-content\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.232897 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpzxz\" (UniqueName: \"kubernetes.io/projected/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-kube-api-access-gpzxz\") pod \"redhat-operators-bwfdk\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.274689 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.339557 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-bhkbs"] Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.341612 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.345784 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.350745 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bhkbs"] Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.352365 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.528929 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-scripts\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.529039 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-config-data\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.529152 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqnjg\" (UniqueName: \"kubernetes.io/projected/693acee8-7f5a-40ce-9b09-57c92bfd97c1-kube-api-access-kqnjg\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.529212 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.631159 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-config-data\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.631976 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqnjg\" (UniqueName: \"kubernetes.io/projected/693acee8-7f5a-40ce-9b09-57c92bfd97c1-kube-api-access-kqnjg\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.632035 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.632066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-scripts\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.638398 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-scripts\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.638755 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-config-data\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.638995 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.648993 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqnjg\" (UniqueName: \"kubernetes.io/projected/693acee8-7f5a-40ce-9b09-57c92bfd97c1-kube-api-access-kqnjg\") pod \"nova-cell1-cell-mapping-bhkbs\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.665531 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:46 crc kubenswrapper[4715]: I1203 22:19:46.891153 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bwfdk"] Dec 03 22:19:47 crc kubenswrapper[4715]: I1203 22:19:47.094285 4715 generic.go:334] "Generic (PLEG): container finished" podID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerID="daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba" exitCode=0 Dec 03 22:19:47 crc kubenswrapper[4715]: I1203 22:19:47.095612 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwfdk" event={"ID":"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6","Type":"ContainerDied","Data":"daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba"} Dec 03 22:19:47 crc kubenswrapper[4715]: I1203 22:19:47.095635 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwfdk" event={"ID":"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6","Type":"ContainerStarted","Data":"6aae528b78716d9c94fb3639afbca5967f55df7682862250ca6497f22012860c"} Dec 03 22:19:47 crc kubenswrapper[4715]: I1203 22:19:47.137515 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-bhkbs"] Dec 03 22:19:47 crc kubenswrapper[4715]: W1203 22:19:47.145930 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod693acee8_7f5a_40ce_9b09_57c92bfd97c1.slice/crio-b9f67392799daaa1369434ed827a717a2fffd82a2125c15687a2a425bc65aaed WatchSource:0}: Error finding container b9f67392799daaa1369434ed827a717a2fffd82a2125c15687a2a425bc65aaed: Status 404 returned error can't find the container with id b9f67392799daaa1369434ed827a717a2fffd82a2125c15687a2a425bc65aaed Dec 03 22:19:47 crc kubenswrapper[4715]: I1203 22:19:47.648456 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3f721fb-a909-4614-8f12-ae64db2c78c9" path="/var/lib/kubelet/pods/b3f721fb-a909-4614-8f12-ae64db2c78c9/volumes" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.112049 4715 generic.go:334] "Generic (PLEG): container finished" podID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerID="b8403e182d586651bdeb70a45cf83dbd31d91f319e6682852d42f4e8610f95fa" exitCode=0 Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.112115 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerDied","Data":"b8403e182d586651bdeb70a45cf83dbd31d91f319e6682852d42f4e8610f95fa"} Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.114577 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bhkbs" event={"ID":"693acee8-7f5a-40ce-9b09-57c92bfd97c1","Type":"ContainerStarted","Data":"bccc200445d5a2106f8a85baa68217d4e6b0dde973a823ea144fcb39d7835e7c"} Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.114637 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bhkbs" event={"ID":"693acee8-7f5a-40ce-9b09-57c92bfd97c1","Type":"ContainerStarted","Data":"b9f67392799daaa1369434ed827a717a2fffd82a2125c15687a2a425bc65aaed"} Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.116548 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwfdk" event={"ID":"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6","Type":"ContainerStarted","Data":"136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf"} Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.145077 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-bhkbs" podStartSLOduration=2.145032865 podStartE2EDuration="2.145032865s" podCreationTimestamp="2025-12-03 22:19:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:19:48.12987964 +0000 UTC m=+2284.872590235" watchObservedRunningTime="2025-12-03 22:19:48.145032865 +0000 UTC m=+2284.887743490" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.214608 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.280660 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpxr4\" (UniqueName: \"kubernetes.io/projected/b35ac264-213f-4e1e-926d-cb6e2b47c684-kube-api-access-dpxr4\") pod \"b35ac264-213f-4e1e-926d-cb6e2b47c684\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.285793 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b35ac264-213f-4e1e-926d-cb6e2b47c684-kube-api-access-dpxr4" (OuterVolumeSpecName: "kube-api-access-dpxr4") pod "b35ac264-213f-4e1e-926d-cb6e2b47c684" (UID: "b35ac264-213f-4e1e-926d-cb6e2b47c684"). InnerVolumeSpecName "kube-api-access-dpxr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.383103 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-scripts\") pod \"b35ac264-213f-4e1e-926d-cb6e2b47c684\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.383184 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-run-httpd\") pod \"b35ac264-213f-4e1e-926d-cb6e2b47c684\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.383381 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-sg-core-conf-yaml\") pod \"b35ac264-213f-4e1e-926d-cb6e2b47c684\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.383793 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-combined-ca-bundle\") pod \"b35ac264-213f-4e1e-926d-cb6e2b47c684\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.383868 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-log-httpd\") pod \"b35ac264-213f-4e1e-926d-cb6e2b47c684\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.383901 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-config-data\") pod \"b35ac264-213f-4e1e-926d-cb6e2b47c684\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.383958 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-ceilometer-tls-certs\") pod \"b35ac264-213f-4e1e-926d-cb6e2b47c684\" (UID: \"b35ac264-213f-4e1e-926d-cb6e2b47c684\") " Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.383783 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b35ac264-213f-4e1e-926d-cb6e2b47c684" (UID: "b35ac264-213f-4e1e-926d-cb6e2b47c684"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.384170 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b35ac264-213f-4e1e-926d-cb6e2b47c684" (UID: "b35ac264-213f-4e1e-926d-cb6e2b47c684"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.384630 4715 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.384658 4715 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b35ac264-213f-4e1e-926d-cb6e2b47c684-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.384672 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpxr4\" (UniqueName: \"kubernetes.io/projected/b35ac264-213f-4e1e-926d-cb6e2b47c684-kube-api-access-dpxr4\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.396854 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-scripts" (OuterVolumeSpecName: "scripts") pod "b35ac264-213f-4e1e-926d-cb6e2b47c684" (UID: "b35ac264-213f-4e1e-926d-cb6e2b47c684"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.419318 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b35ac264-213f-4e1e-926d-cb6e2b47c684" (UID: "b35ac264-213f-4e1e-926d-cb6e2b47c684"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.471871 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b35ac264-213f-4e1e-926d-cb6e2b47c684" (UID: "b35ac264-213f-4e1e-926d-cb6e2b47c684"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.472707 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b35ac264-213f-4e1e-926d-cb6e2b47c684" (UID: "b35ac264-213f-4e1e-926d-cb6e2b47c684"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.486614 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.486702 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.486712 4715 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.486721 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.543644 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-config-data" (OuterVolumeSpecName: "config-data") pod "b35ac264-213f-4e1e-926d-cb6e2b47c684" (UID: "b35ac264-213f-4e1e-926d-cb6e2b47c684"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:48 crc kubenswrapper[4715]: I1203 22:19:48.587560 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b35ac264-213f-4e1e-926d-cb6e2b47c684-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.127556 4715 generic.go:334] "Generic (PLEG): container finished" podID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerID="136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf" exitCode=0 Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.127638 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwfdk" event={"ID":"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6","Type":"ContainerDied","Data":"136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf"} Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.141893 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.142428 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b35ac264-213f-4e1e-926d-cb6e2b47c684","Type":"ContainerDied","Data":"dfb7e742e29f6074dd1fa35623c2e33f751b24d058067f36b73ed433d08b675e"} Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.142494 4715 scope.go:117] "RemoveContainer" containerID="1abc61a4c85ca4fbc980e3269d44a353e556e8a42133803c495d40fa5bc6a9b2" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.166738 4715 scope.go:117] "RemoveContainer" containerID="86b35addd818249ba60715783d48f08c447ef768a6c26e5d6bf03bfa876b5479" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.207903 4715 scope.go:117] "RemoveContainer" containerID="dab8336003aa53b24b0cf172c6b7088b96dc670fb9b74a66819a0c0367c188f4" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.217278 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.230934 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.239340 4715 scope.go:117] "RemoveContainer" containerID="b8403e182d586651bdeb70a45cf83dbd31d91f319e6682852d42f4e8610f95fa" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.243557 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:49 crc kubenswrapper[4715]: E1203 22:19:49.243962 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="ceilometer-notification-agent" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.243980 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="ceilometer-notification-agent" Dec 03 22:19:49 crc kubenswrapper[4715]: E1203 22:19:49.244005 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="proxy-httpd" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.244012 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="proxy-httpd" Dec 03 22:19:49 crc kubenswrapper[4715]: E1203 22:19:49.244026 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="sg-core" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.244032 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="sg-core" Dec 03 22:19:49 crc kubenswrapper[4715]: E1203 22:19:49.244058 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="ceilometer-central-agent" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.244064 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="ceilometer-central-agent" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.244232 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="proxy-httpd" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.244242 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="ceilometer-notification-agent" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.244256 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="ceilometer-central-agent" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.244276 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" containerName="sg-core" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.245888 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.250301 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.250489 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.254053 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.270568 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.401659 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-config-data\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.402940 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmlw9\" (UniqueName: \"kubernetes.io/projected/ac3428ad-2684-4ca1-ab1e-254306dc84a3-kube-api-access-dmlw9\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.403144 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac3428ad-2684-4ca1-ab1e-254306dc84a3-run-httpd\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.403226 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.403437 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac3428ad-2684-4ca1-ab1e-254306dc84a3-log-httpd\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.403537 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.403735 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.403790 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-scripts\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505135 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac3428ad-2684-4ca1-ab1e-254306dc84a3-run-httpd\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505413 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505476 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac3428ad-2684-4ca1-ab1e-254306dc84a3-log-httpd\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505530 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505637 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-scripts\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505668 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-config-data\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505693 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmlw9\" (UniqueName: \"kubernetes.io/projected/ac3428ad-2684-4ca1-ab1e-254306dc84a3-kube-api-access-dmlw9\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.505799 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac3428ad-2684-4ca1-ab1e-254306dc84a3-run-httpd\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.506519 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac3428ad-2684-4ca1-ab1e-254306dc84a3-log-httpd\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.513732 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-scripts\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.513735 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.513840 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.513851 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.514543 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac3428ad-2684-4ca1-ab1e-254306dc84a3-config-data\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.522763 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmlw9\" (UniqueName: \"kubernetes.io/projected/ac3428ad-2684-4ca1-ab1e-254306dc84a3-kube-api-access-dmlw9\") pod \"ceilometer-0\" (UID: \"ac3428ad-2684-4ca1-ab1e-254306dc84a3\") " pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.578971 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 22:19:49 crc kubenswrapper[4715]: I1203 22:19:49.651139 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b35ac264-213f-4e1e-926d-cb6e2b47c684" path="/var/lib/kubelet/pods/b35ac264-213f-4e1e-926d-cb6e2b47c684/volumes" Dec 03 22:19:50 crc kubenswrapper[4715]: W1203 22:19:50.437832 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac3428ad_2684_4ca1_ab1e_254306dc84a3.slice/crio-47dc17f5ae22ff0854c6781baf0821f9e5f24b5b0e7f169815493ac7750bc1c8 WatchSource:0}: Error finding container 47dc17f5ae22ff0854c6781baf0821f9e5f24b5b0e7f169815493ac7750bc1c8: Status 404 returned error can't find the container with id 47dc17f5ae22ff0854c6781baf0821f9e5f24b5b0e7f169815493ac7750bc1c8 Dec 03 22:19:50 crc kubenswrapper[4715]: I1203 22:19:50.442938 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 22:19:51 crc kubenswrapper[4715]: I1203 22:19:51.191107 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwfdk" event={"ID":"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6","Type":"ContainerStarted","Data":"2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c"} Dec 03 22:19:51 crc kubenswrapper[4715]: I1203 22:19:51.193277 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac3428ad-2684-4ca1-ab1e-254306dc84a3","Type":"ContainerStarted","Data":"7d63f1861828ebe27668d4cce1f84223e322538a6eb619a82a6dc0cc7e3b3f5e"} Dec 03 22:19:51 crc kubenswrapper[4715]: I1203 22:19:51.193308 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac3428ad-2684-4ca1-ab1e-254306dc84a3","Type":"ContainerStarted","Data":"47dc17f5ae22ff0854c6781baf0821f9e5f24b5b0e7f169815493ac7750bc1c8"} Dec 03 22:19:51 crc kubenswrapper[4715]: I1203 22:19:51.223895 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bwfdk" podStartSLOduration=3.401839132 podStartE2EDuration="6.223878321s" podCreationTimestamp="2025-12-03 22:19:45 +0000 UTC" firstStartedPulling="2025-12-03 22:19:47.096216437 +0000 UTC m=+2283.838927032" lastFinishedPulling="2025-12-03 22:19:49.918255626 +0000 UTC m=+2286.660966221" observedRunningTime="2025-12-03 22:19:51.214871441 +0000 UTC m=+2287.957582036" watchObservedRunningTime="2025-12-03 22:19:51.223878321 +0000 UTC m=+2287.966588916" Dec 03 22:19:51 crc kubenswrapper[4715]: I1203 22:19:51.420088 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 22:19:51 crc kubenswrapper[4715]: I1203 22:19:51.420489 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 22:19:52 crc kubenswrapper[4715]: I1203 22:19:52.435830 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 22:19:52 crc kubenswrapper[4715]: I1203 22:19:52.435921 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.197:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 22:19:55 crc kubenswrapper[4715]: I1203 22:19:55.244720 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac3428ad-2684-4ca1-ab1e-254306dc84a3","Type":"ContainerStarted","Data":"c35aba152aa7ecb23b37bf6e5bb17fc9f9b28a48e58797097fd903fdfa090c8f"} Dec 03 22:19:55 crc kubenswrapper[4715]: I1203 22:19:55.247868 4715 generic.go:334] "Generic (PLEG): container finished" podID="693acee8-7f5a-40ce-9b09-57c92bfd97c1" containerID="bccc200445d5a2106f8a85baa68217d4e6b0dde973a823ea144fcb39d7835e7c" exitCode=0 Dec 03 22:19:55 crc kubenswrapper[4715]: I1203 22:19:55.247913 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bhkbs" event={"ID":"693acee8-7f5a-40ce-9b09-57c92bfd97c1","Type":"ContainerDied","Data":"bccc200445d5a2106f8a85baa68217d4e6b0dde973a823ea144fcb39d7835e7c"} Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.258291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac3428ad-2684-4ca1-ab1e-254306dc84a3","Type":"ContainerStarted","Data":"1a1c1a3f1c50e92f25901c1c6d97a8b34301bbfa592cdfae32ea816b76942a0e"} Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.275451 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.275536 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.329757 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.634458 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:19:56 crc kubenswrapper[4715]: E1203 22:19:56.635050 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.668628 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.769299 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-combined-ca-bundle\") pod \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.769657 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-config-data\") pod \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.769687 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-scripts\") pod \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.769716 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqnjg\" (UniqueName: \"kubernetes.io/projected/693acee8-7f5a-40ce-9b09-57c92bfd97c1-kube-api-access-kqnjg\") pod \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\" (UID: \"693acee8-7f5a-40ce-9b09-57c92bfd97c1\") " Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.778685 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/693acee8-7f5a-40ce-9b09-57c92bfd97c1-kube-api-access-kqnjg" (OuterVolumeSpecName: "kube-api-access-kqnjg") pod "693acee8-7f5a-40ce-9b09-57c92bfd97c1" (UID: "693acee8-7f5a-40ce-9b09-57c92bfd97c1"). InnerVolumeSpecName "kube-api-access-kqnjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.781924 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-scripts" (OuterVolumeSpecName: "scripts") pod "693acee8-7f5a-40ce-9b09-57c92bfd97c1" (UID: "693acee8-7f5a-40ce-9b09-57c92bfd97c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.819799 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "693acee8-7f5a-40ce-9b09-57c92bfd97c1" (UID: "693acee8-7f5a-40ce-9b09-57c92bfd97c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.827129 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-config-data" (OuterVolumeSpecName: "config-data") pod "693acee8-7f5a-40ce-9b09-57c92bfd97c1" (UID: "693acee8-7f5a-40ce-9b09-57c92bfd97c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.872909 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.872938 4715 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.872949 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqnjg\" (UniqueName: \"kubernetes.io/projected/693acee8-7f5a-40ce-9b09-57c92bfd97c1-kube-api-access-kqnjg\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:56 crc kubenswrapper[4715]: I1203 22:19:56.872959 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/693acee8-7f5a-40ce-9b09-57c92bfd97c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.304824 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-bhkbs" Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.305087 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-bhkbs" event={"ID":"693acee8-7f5a-40ce-9b09-57c92bfd97c1","Type":"ContainerDied","Data":"b9f67392799daaa1369434ed827a717a2fffd82a2125c15687a2a425bc65aaed"} Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.305340 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9f67392799daaa1369434ed827a717a2fffd82a2125c15687a2a425bc65aaed" Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.350635 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.428974 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bwfdk"] Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.499728 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.500031 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-log" containerID="cri-o://054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86" gracePeriod=30 Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.500618 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-api" containerID="cri-o://8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2" gracePeriod=30 Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.512998 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.513215 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="f7481a5f-d565-4ac9-b629-bfb792cb54f2" containerName="nova-scheduler-scheduler" containerID="cri-o://c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227" gracePeriod=30 Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.551492 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.551868 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-log" containerID="cri-o://8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f" gracePeriod=30 Dec 03 22:19:57 crc kubenswrapper[4715]: I1203 22:19:57.552205 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-metadata" containerID="cri-o://d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2" gracePeriod=30 Dec 03 22:19:58 crc kubenswrapper[4715]: I1203 22:19:58.313953 4715 generic.go:334] "Generic (PLEG): container finished" podID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerID="054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86" exitCode=143 Dec 03 22:19:58 crc kubenswrapper[4715]: I1203 22:19:58.314063 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45e6fe90-4116-4017-bd10-3f3b42e68c3e","Type":"ContainerDied","Data":"054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86"} Dec 03 22:19:58 crc kubenswrapper[4715]: I1203 22:19:58.317215 4715 generic.go:334] "Generic (PLEG): container finished" podID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerID="8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f" exitCode=143 Dec 03 22:19:58 crc kubenswrapper[4715]: I1203 22:19:58.317297 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2533a7c4-9e84-4f23-b2d5-065ca474fbb4","Type":"ContainerDied","Data":"8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f"} Dec 03 22:19:58 crc kubenswrapper[4715]: I1203 22:19:58.320059 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac3428ad-2684-4ca1-ab1e-254306dc84a3","Type":"ContainerStarted","Data":"b6b49434a61cba06b531527857c55798c1f80329ce953355d80bb79ef5569faf"} Dec 03 22:19:58 crc kubenswrapper[4715]: I1203 22:19:58.320435 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 22:19:58 crc kubenswrapper[4715]: I1203 22:19:58.347141 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.597839673 podStartE2EDuration="9.347121406s" podCreationTimestamp="2025-12-03 22:19:49 +0000 UTC" firstStartedPulling="2025-12-03 22:19:50.443280567 +0000 UTC m=+2287.185991172" lastFinishedPulling="2025-12-03 22:19:57.19256231 +0000 UTC m=+2293.935272905" observedRunningTime="2025-12-03 22:19:58.341570157 +0000 UTC m=+2295.084280772" watchObservedRunningTime="2025-12-03 22:19:58.347121406 +0000 UTC m=+2295.089832001" Dec 03 22:19:59 crc kubenswrapper[4715]: I1203 22:19:59.331286 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bwfdk" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerName="registry-server" containerID="cri-o://2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c" gracePeriod=2 Dec 03 22:19:59 crc kubenswrapper[4715]: I1203 22:19:59.897878 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:19:59 crc kubenswrapper[4715]: I1203 22:19:59.941526 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpzxz\" (UniqueName: \"kubernetes.io/projected/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-kube-api-access-gpzxz\") pod \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " Dec 03 22:19:59 crc kubenswrapper[4715]: I1203 22:19:59.941667 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-catalog-content\") pod \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " Dec 03 22:19:59 crc kubenswrapper[4715]: I1203 22:19:59.941780 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-utilities\") pod \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\" (UID: \"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6\") " Dec 03 22:19:59 crc kubenswrapper[4715]: I1203 22:19:59.942986 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-utilities" (OuterVolumeSpecName: "utilities") pod "63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" (UID: "63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:19:59 crc kubenswrapper[4715]: I1203 22:19:59.949647 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-kube-api-access-gpzxz" (OuterVolumeSpecName: "kube-api-access-gpzxz") pod "63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" (UID: "63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6"). InnerVolumeSpecName "kube-api-access-gpzxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.044074 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpzxz\" (UniqueName: \"kubernetes.io/projected/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-kube-api-access-gpzxz\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.044111 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.048876 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" (UID: "63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.146998 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.347529 4715 generic.go:334] "Generic (PLEG): container finished" podID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerID="2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c" exitCode=0 Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.347614 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bwfdk" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.347630 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwfdk" event={"ID":"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6","Type":"ContainerDied","Data":"2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c"} Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.347717 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bwfdk" event={"ID":"63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6","Type":"ContainerDied","Data":"6aae528b78716d9c94fb3639afbca5967f55df7682862250ca6497f22012860c"} Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.347749 4715 scope.go:117] "RemoveContainer" containerID="2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.390059 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bwfdk"] Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.393989 4715 scope.go:117] "RemoveContainer" containerID="136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.399933 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bwfdk"] Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.420368 4715 scope.go:117] "RemoveContainer" containerID="daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.462339 4715 scope.go:117] "RemoveContainer" containerID="2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c" Dec 03 22:20:00 crc kubenswrapper[4715]: E1203 22:20:00.462818 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c\": container with ID starting with 2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c not found: ID does not exist" containerID="2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.462852 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c"} err="failed to get container status \"2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c\": rpc error: code = NotFound desc = could not find container \"2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c\": container with ID starting with 2d33a8c130d30c483ad865bce2813715318e4f1d17d021b252b0f176a6f9c72c not found: ID does not exist" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.462887 4715 scope.go:117] "RemoveContainer" containerID="136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf" Dec 03 22:20:00 crc kubenswrapper[4715]: E1203 22:20:00.463292 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf\": container with ID starting with 136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf not found: ID does not exist" containerID="136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.463357 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf"} err="failed to get container status \"136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf\": rpc error: code = NotFound desc = could not find container \"136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf\": container with ID starting with 136ddce56b53933e23fb405dfb717bc9eb0a16f87a7d142af96c482072dc22bf not found: ID does not exist" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.463406 4715 scope.go:117] "RemoveContainer" containerID="daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba" Dec 03 22:20:00 crc kubenswrapper[4715]: E1203 22:20:00.463766 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba\": container with ID starting with daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba not found: ID does not exist" containerID="daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.463800 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba"} err="failed to get container status \"daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba\": rpc error: code = NotFound desc = could not find container \"daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba\": container with ID starting with daba585fd24c82acc4f0e5d14f23a4006ea88dd0081679cac8894eb972731aba not found: ID does not exist" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.722737 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:54902->10.217.0.190:8775: read: connection reset by peer" Dec 03 22:20:00 crc kubenswrapper[4715]: I1203 22:20:00.722830 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.190:8775/\": read tcp 10.217.0.2:54900->10.217.0.190:8775: read: connection reset by peer" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.136125 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.171337 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-combined-ca-bundle\") pod \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.171490 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw6hs\" (UniqueName: \"kubernetes.io/projected/45e6fe90-4116-4017-bd10-3f3b42e68c3e-kube-api-access-bw6hs\") pod \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.171544 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-public-tls-certs\") pod \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.171584 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-internal-tls-certs\") pod \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.171670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-config-data\") pod \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.171733 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e6fe90-4116-4017-bd10-3f3b42e68c3e-logs\") pod \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\" (UID: \"45e6fe90-4116-4017-bd10-3f3b42e68c3e\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.173055 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45e6fe90-4116-4017-bd10-3f3b42e68c3e-logs" (OuterVolumeSpecName: "logs") pod "45e6fe90-4116-4017-bd10-3f3b42e68c3e" (UID: "45e6fe90-4116-4017-bd10-3f3b42e68c3e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.177307 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45e6fe90-4116-4017-bd10-3f3b42e68c3e-kube-api-access-bw6hs" (OuterVolumeSpecName: "kube-api-access-bw6hs") pod "45e6fe90-4116-4017-bd10-3f3b42e68c3e" (UID: "45e6fe90-4116-4017-bd10-3f3b42e68c3e"). InnerVolumeSpecName "kube-api-access-bw6hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.184703 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.230889 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-config-data" (OuterVolumeSpecName: "config-data") pod "45e6fe90-4116-4017-bd10-3f3b42e68c3e" (UID: "45e6fe90-4116-4017-bd10-3f3b42e68c3e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.240238 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45e6fe90-4116-4017-bd10-3f3b42e68c3e" (UID: "45e6fe90-4116-4017-bd10-3f3b42e68c3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.258569 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "45e6fe90-4116-4017-bd10-3f3b42e68c3e" (UID: "45e6fe90-4116-4017-bd10-3f3b42e68c3e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.259751 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "45e6fe90-4116-4017-bd10-3f3b42e68c3e" (UID: "45e6fe90-4116-4017-bd10-3f3b42e68c3e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.273010 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-logs\") pod \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.273377 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-nova-metadata-tls-certs\") pod \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.273403 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-combined-ca-bundle\") pod \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.273443 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-config-data\") pod \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.273617 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgl5p\" (UniqueName: \"kubernetes.io/projected/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-kube-api-access-rgl5p\") pod \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\" (UID: \"2533a7c4-9e84-4f23-b2d5-065ca474fbb4\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.273983 4715 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.274002 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.274011 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45e6fe90-4116-4017-bd10-3f3b42e68c3e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.274019 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.274030 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw6hs\" (UniqueName: \"kubernetes.io/projected/45e6fe90-4116-4017-bd10-3f3b42e68c3e-kube-api-access-bw6hs\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.274039 4715 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45e6fe90-4116-4017-bd10-3f3b42e68c3e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.274426 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-logs" (OuterVolumeSpecName: "logs") pod "2533a7c4-9e84-4f23-b2d5-065ca474fbb4" (UID: "2533a7c4-9e84-4f23-b2d5-065ca474fbb4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.279880 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-kube-api-access-rgl5p" (OuterVolumeSpecName: "kube-api-access-rgl5p") pod "2533a7c4-9e84-4f23-b2d5-065ca474fbb4" (UID: "2533a7c4-9e84-4f23-b2d5-065ca474fbb4"). InnerVolumeSpecName "kube-api-access-rgl5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.305484 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-config-data" (OuterVolumeSpecName: "config-data") pod "2533a7c4-9e84-4f23-b2d5-065ca474fbb4" (UID: "2533a7c4-9e84-4f23-b2d5-065ca474fbb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.310434 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2533a7c4-9e84-4f23-b2d5-065ca474fbb4" (UID: "2533a7c4-9e84-4f23-b2d5-065ca474fbb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.327047 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.356537 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "2533a7c4-9e84-4f23-b2d5-065ca474fbb4" (UID: "2533a7c4-9e84-4f23-b2d5-065ca474fbb4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.378034 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-combined-ca-bundle\") pod \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.378358 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n42d5\" (UniqueName: \"kubernetes.io/projected/f7481a5f-d565-4ac9-b629-bfb792cb54f2-kube-api-access-n42d5\") pod \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.378572 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-config-data\") pod \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\" (UID: \"f7481a5f-d565-4ac9-b629-bfb792cb54f2\") " Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.382825 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgl5p\" (UniqueName: \"kubernetes.io/projected/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-kube-api-access-rgl5p\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.382916 4715 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-logs\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.382984 4715 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.383070 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.383128 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2533a7c4-9e84-4f23-b2d5-065ca474fbb4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.398616 4715 generic.go:334] "Generic (PLEG): container finished" podID="f7481a5f-d565-4ac9-b629-bfb792cb54f2" containerID="c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227" exitCode=0 Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.398809 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.398861 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7481a5f-d565-4ac9-b629-bfb792cb54f2","Type":"ContainerDied","Data":"c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227"} Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.398914 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"f7481a5f-d565-4ac9-b629-bfb792cb54f2","Type":"ContainerDied","Data":"064a1d5770824c6acc1a08fd16fe5aaa6a7be27072b1c498d784b9506ba5e91b"} Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.398935 4715 scope.go:117] "RemoveContainer" containerID="c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.404061 4715 generic.go:334] "Generic (PLEG): container finished" podID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerID="8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2" exitCode=0 Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.404144 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45e6fe90-4116-4017-bd10-3f3b42e68c3e","Type":"ContainerDied","Data":"8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2"} Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.404165 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"45e6fe90-4116-4017-bd10-3f3b42e68c3e","Type":"ContainerDied","Data":"9c77f7dfc17a7c2268e08385dcd03f7bdaba27440acefc9c741b856c2460f710"} Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.404167 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.405629 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7481a5f-d565-4ac9-b629-bfb792cb54f2-kube-api-access-n42d5" (OuterVolumeSpecName: "kube-api-access-n42d5") pod "f7481a5f-d565-4ac9-b629-bfb792cb54f2" (UID: "f7481a5f-d565-4ac9-b629-bfb792cb54f2"). InnerVolumeSpecName "kube-api-access-n42d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.410199 4715 generic.go:334] "Generic (PLEG): container finished" podID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerID="d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2" exitCode=0 Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.410381 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.410517 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2533a7c4-9e84-4f23-b2d5-065ca474fbb4","Type":"ContainerDied","Data":"d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2"} Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.410657 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2533a7c4-9e84-4f23-b2d5-065ca474fbb4","Type":"ContainerDied","Data":"810c881ec3f811504782179593158c21ae7f2e54316940ee414680cfb993cf4b"} Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.440254 4715 scope.go:117] "RemoveContainer" containerID="c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.445816 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227\": container with ID starting with c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227 not found: ID does not exist" containerID="c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.445882 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227"} err="failed to get container status \"c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227\": rpc error: code = NotFound desc = could not find container \"c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227\": container with ID starting with c4fd11d86899d43b7d918b65eaeea9518bb31580c526a43d80e2d2592ae2d227 not found: ID does not exist" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.445919 4715 scope.go:117] "RemoveContainer" containerID="8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.445800 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-config-data" (OuterVolumeSpecName: "config-data") pod "f7481a5f-d565-4ac9-b629-bfb792cb54f2" (UID: "f7481a5f-d565-4ac9-b629-bfb792cb54f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.465029 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.468675 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7481a5f-d565-4ac9-b629-bfb792cb54f2" (UID: "f7481a5f-d565-4ac9-b629-bfb792cb54f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.475736 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.484480 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.484522 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7481a5f-d565-4ac9-b629-bfb792cb54f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.484540 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n42d5\" (UniqueName: \"kubernetes.io/projected/f7481a5f-d565-4ac9-b629-bfb792cb54f2-kube-api-access-n42d5\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.491624 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.499881 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.509836 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510257 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-log" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510274 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-log" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510288 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-metadata" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510295 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-metadata" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510308 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerName="extract-utilities" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510314 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerName="extract-utilities" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510323 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerName="registry-server" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510329 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerName="registry-server" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510341 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="693acee8-7f5a-40ce-9b09-57c92bfd97c1" containerName="nova-manage" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510346 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="693acee8-7f5a-40ce-9b09-57c92bfd97c1" containerName="nova-manage" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510363 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7481a5f-d565-4ac9-b629-bfb792cb54f2" containerName="nova-scheduler-scheduler" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510369 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7481a5f-d565-4ac9-b629-bfb792cb54f2" containerName="nova-scheduler-scheduler" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510375 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-api" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510381 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-api" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510394 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerName="extract-content" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510400 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerName="extract-content" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.510408 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-log" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510413 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-log" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510599 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" containerName="registry-server" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510613 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-api" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510621 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-metadata" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510633 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" containerName="nova-metadata-log" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510651 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7481a5f-d565-4ac9-b629-bfb792cb54f2" containerName="nova-scheduler-scheduler" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510664 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" containerName="nova-api-log" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.510669 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="693acee8-7f5a-40ce-9b09-57c92bfd97c1" containerName="nova-manage" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.511634 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.513755 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.514010 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.523595 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.525866 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.530549 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.530868 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.531002 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.532579 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.542602 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.582928 4715 scope.go:117] "RemoveContainer" containerID="054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586568 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586626 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586660 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwqfs\" (UniqueName: \"kubernetes.io/projected/b079ad3a-bcab-48cc-99da-f7f656195558-kube-api-access-fwqfs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586709 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5spk\" (UniqueName: \"kubernetes.io/projected/0a252fdb-1b30-4278-8bfc-471881a27f21-kube-api-access-l5spk\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586741 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-config-data\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586788 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-public-tls-certs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586815 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b079ad3a-bcab-48cc-99da-f7f656195558-logs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586840 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a252fdb-1b30-4278-8bfc-471881a27f21-logs\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586877 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-config-data\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586904 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.586930 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.609923 4715 scope.go:117] "RemoveContainer" containerID="8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.610628 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2\": container with ID starting with 8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2 not found: ID does not exist" containerID="8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.610670 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2"} err="failed to get container status \"8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2\": rpc error: code = NotFound desc = could not find container \"8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2\": container with ID starting with 8ddef36cfe54088cb1264068148651b3cd84bb8882167fbdda5d389e07fd80f2 not found: ID does not exist" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.610699 4715 scope.go:117] "RemoveContainer" containerID="054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.611663 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86\": container with ID starting with 054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86 not found: ID does not exist" containerID="054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.611737 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86"} err="failed to get container status \"054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86\": rpc error: code = NotFound desc = could not find container \"054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86\": container with ID starting with 054296aa954402f8c4aea03ce678de8350e02c421988df9a06b88a3a70d83e86 not found: ID does not exist" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.611780 4715 scope.go:117] "RemoveContainer" containerID="d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.630876 4715 scope.go:117] "RemoveContainer" containerID="8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.648944 4715 scope.go:117] "RemoveContainer" containerID="d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.651418 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2533a7c4-9e84-4f23-b2d5-065ca474fbb4" path="/var/lib/kubelet/pods/2533a7c4-9e84-4f23-b2d5-065ca474fbb4/volumes" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.651720 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2\": container with ID starting with d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2 not found: ID does not exist" containerID="d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.651780 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2"} err="failed to get container status \"d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2\": rpc error: code = NotFound desc = could not find container \"d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2\": container with ID starting with d51aa2d965288eb5fba7fc6afd09bb21c42551b8370a36d1b1decb1404f0dba2 not found: ID does not exist" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.651815 4715 scope.go:117] "RemoveContainer" containerID="8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.652192 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45e6fe90-4116-4017-bd10-3f3b42e68c3e" path="/var/lib/kubelet/pods/45e6fe90-4116-4017-bd10-3f3b42e68c3e/volumes" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.652860 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6" path="/var/lib/kubelet/pods/63381f76-d4d1-4bb0-9d2d-bdf2f9a2b1b6/volumes" Dec 03 22:20:01 crc kubenswrapper[4715]: E1203 22:20:01.653862 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f\": container with ID starting with 8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f not found: ID does not exist" containerID="8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.653891 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f"} err="failed to get container status \"8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f\": rpc error: code = NotFound desc = could not find container \"8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f\": container with ID starting with 8bc0ae1a87bb7a54ce278117f8446ec66f3eb59a1b51d59c5b086ce3a8ee8e1f not found: ID does not exist" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.689118 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.689234 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.689275 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwqfs\" (UniqueName: \"kubernetes.io/projected/b079ad3a-bcab-48cc-99da-f7f656195558-kube-api-access-fwqfs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.689374 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5spk\" (UniqueName: \"kubernetes.io/projected/0a252fdb-1b30-4278-8bfc-471881a27f21-kube-api-access-l5spk\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.690424 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-config-data\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.691941 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-public-tls-certs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.692007 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b079ad3a-bcab-48cc-99da-f7f656195558-logs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.692036 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a252fdb-1b30-4278-8bfc-471881a27f21-logs\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.692112 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-config-data\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.692141 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.692193 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.692474 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b079ad3a-bcab-48cc-99da-f7f656195558-logs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.692960 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.693054 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a252fdb-1b30-4278-8bfc-471881a27f21-logs\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.695159 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.696338 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-internal-tls-certs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.698271 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-config-data\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.702524 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-config-data\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.702924 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a252fdb-1b30-4278-8bfc-471881a27f21-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.703853 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b079ad3a-bcab-48cc-99da-f7f656195558-public-tls-certs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.710972 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5spk\" (UniqueName: \"kubernetes.io/projected/0a252fdb-1b30-4278-8bfc-471881a27f21-kube-api-access-l5spk\") pod \"nova-metadata-0\" (UID: \"0a252fdb-1b30-4278-8bfc-471881a27f21\") " pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.712925 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwqfs\" (UniqueName: \"kubernetes.io/projected/b079ad3a-bcab-48cc-99da-f7f656195558-kube-api-access-fwqfs\") pod \"nova-api-0\" (UID: \"b079ad3a-bcab-48cc-99da-f7f656195558\") " pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.800844 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.816279 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.830645 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.831971 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.834199 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.847332 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.896685 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e79bc96-17dc-43d2-b049-da6de097be5e-config-data\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.896753 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e79bc96-17dc-43d2-b049-da6de097be5e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.896898 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcdln\" (UniqueName: \"kubernetes.io/projected/5e79bc96-17dc-43d2-b049-da6de097be5e-kube-api-access-kcdln\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.897589 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.908376 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.998611 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e79bc96-17dc-43d2-b049-da6de097be5e-config-data\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.998700 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e79bc96-17dc-43d2-b049-da6de097be5e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:01 crc kubenswrapper[4715]: I1203 22:20:01.998760 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcdln\" (UniqueName: \"kubernetes.io/projected/5e79bc96-17dc-43d2-b049-da6de097be5e-kube-api-access-kcdln\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:02 crc kubenswrapper[4715]: I1203 22:20:02.004148 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e79bc96-17dc-43d2-b049-da6de097be5e-config-data\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:02 crc kubenswrapper[4715]: I1203 22:20:02.007492 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e79bc96-17dc-43d2-b049-da6de097be5e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:02 crc kubenswrapper[4715]: I1203 22:20:02.020831 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcdln\" (UniqueName: \"kubernetes.io/projected/5e79bc96-17dc-43d2-b049-da6de097be5e-kube-api-access-kcdln\") pod \"nova-scheduler-0\" (UID: \"5e79bc96-17dc-43d2-b049-da6de097be5e\") " pod="openstack/nova-scheduler-0" Dec 03 22:20:02 crc kubenswrapper[4715]: I1203 22:20:02.150373 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 22:20:02 crc kubenswrapper[4715]: I1203 22:20:02.402030 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 22:20:02 crc kubenswrapper[4715]: I1203 22:20:02.426729 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b079ad3a-bcab-48cc-99da-f7f656195558","Type":"ContainerStarted","Data":"c5381223ddb3ba373cd8531f8e0d4508d5ca52d4bbae02334e5f50933737b77c"} Dec 03 22:20:02 crc kubenswrapper[4715]: I1203 22:20:02.443112 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 22:20:02 crc kubenswrapper[4715]: I1203 22:20:02.604776 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 22:20:02 crc kubenswrapper[4715]: W1203 22:20:02.618541 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e79bc96_17dc_43d2_b049_da6de097be5e.slice/crio-807319ba2d0a72dc89410880b342306a8c38045d80b43d56ccc7387ae120d784 WatchSource:0}: Error finding container 807319ba2d0a72dc89410880b342306a8c38045d80b43d56ccc7387ae120d784: Status 404 returned error can't find the container with id 807319ba2d0a72dc89410880b342306a8c38045d80b43d56ccc7387ae120d784 Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.446311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b079ad3a-bcab-48cc-99da-f7f656195558","Type":"ContainerStarted","Data":"0e486bbc246de79d9b762fd3076d2a2231b2829a7525af87b8ee3c2f8308e66f"} Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.446416 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"b079ad3a-bcab-48cc-99da-f7f656195558","Type":"ContainerStarted","Data":"e90de91dca8ecdfcf0cfdc209ad08018b8aabb59baf33f2ab0aaf6466efbb674"} Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.448922 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5e79bc96-17dc-43d2-b049-da6de097be5e","Type":"ContainerStarted","Data":"124a0a416f73155d08034e04db60877f93e54672e403cc81199083655f41502b"} Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.448992 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"5e79bc96-17dc-43d2-b049-da6de097be5e","Type":"ContainerStarted","Data":"807319ba2d0a72dc89410880b342306a8c38045d80b43d56ccc7387ae120d784"} Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.452195 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a252fdb-1b30-4278-8bfc-471881a27f21","Type":"ContainerStarted","Data":"e317f91e8f2bc31a062c53ef8f5f344d2372814618f9ad827e511c9fd8b4bc94"} Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.452243 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a252fdb-1b30-4278-8bfc-471881a27f21","Type":"ContainerStarted","Data":"a2cf8219f227863942cd503d822668183252184eb6ddae637623e98f034bc2b8"} Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.452263 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0a252fdb-1b30-4278-8bfc-471881a27f21","Type":"ContainerStarted","Data":"d29241f3058dd680d3a57e8419457399f60c8be36a7596bcf4b556e75f55d08a"} Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.522894 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.522873728 podStartE2EDuration="2.522873728s" podCreationTimestamp="2025-12-03 22:20:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:20:03.501968569 +0000 UTC m=+2300.244679164" watchObservedRunningTime="2025-12-03 22:20:03.522873728 +0000 UTC m=+2300.265584323" Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.562744 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.562723274 podStartE2EDuration="2.562723274s" podCreationTimestamp="2025-12-03 22:20:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:20:03.530971985 +0000 UTC m=+2300.273682590" watchObservedRunningTime="2025-12-03 22:20:03.562723274 +0000 UTC m=+2300.305433879" Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.577898 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.577872579 podStartE2EDuration="2.577872579s" podCreationTimestamp="2025-12-03 22:20:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:20:03.574198191 +0000 UTC m=+2300.316908786" watchObservedRunningTime="2025-12-03 22:20:03.577872579 +0000 UTC m=+2300.320583174" Dec 03 22:20:03 crc kubenswrapper[4715]: I1203 22:20:03.659279 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7481a5f-d565-4ac9-b629-bfb792cb54f2" path="/var/lib/kubelet/pods/f7481a5f-d565-4ac9-b629-bfb792cb54f2/volumes" Dec 03 22:20:06 crc kubenswrapper[4715]: I1203 22:20:06.898266 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 22:20:06 crc kubenswrapper[4715]: I1203 22:20:06.898925 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 22:20:07 crc kubenswrapper[4715]: I1203 22:20:07.151043 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 22:20:09 crc kubenswrapper[4715]: I1203 22:20:09.634560 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:20:09 crc kubenswrapper[4715]: E1203 22:20:09.635303 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:20:11 crc kubenswrapper[4715]: I1203 22:20:11.898714 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 22:20:11 crc kubenswrapper[4715]: I1203 22:20:11.899087 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 22:20:11 crc kubenswrapper[4715]: I1203 22:20:11.909726 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 22:20:11 crc kubenswrapper[4715]: I1203 22:20:11.909771 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 22:20:12 crc kubenswrapper[4715]: I1203 22:20:12.151373 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 22:20:12 crc kubenswrapper[4715]: I1203 22:20:12.179083 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 22:20:12 crc kubenswrapper[4715]: I1203 22:20:12.579192 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 22:20:12 crc kubenswrapper[4715]: I1203 22:20:12.952822 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0a252fdb-1b30-4278-8bfc-471881a27f21" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 22:20:12 crc kubenswrapper[4715]: I1203 22:20:12.952898 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b079ad3a-bcab-48cc-99da-f7f656195558" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 22:20:12 crc kubenswrapper[4715]: I1203 22:20:12.952941 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0a252fdb-1b30-4278-8bfc-471881a27f21" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.201:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 22:20:12 crc kubenswrapper[4715]: I1203 22:20:12.952995 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="b079ad3a-bcab-48cc-99da-f7f656195558" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 22:20:19 crc kubenswrapper[4715]: I1203 22:20:19.595490 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 22:20:21 crc kubenswrapper[4715]: I1203 22:20:21.904772 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 22:20:21 crc kubenswrapper[4715]: I1203 22:20:21.911639 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 22:20:21 crc kubenswrapper[4715]: I1203 22:20:21.914290 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 22:20:21 crc kubenswrapper[4715]: I1203 22:20:21.922041 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 22:20:21 crc kubenswrapper[4715]: I1203 22:20:21.922603 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 22:20:21 crc kubenswrapper[4715]: I1203 22:20:21.923938 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 22:20:21 crc kubenswrapper[4715]: I1203 22:20:21.938975 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 22:20:22 crc kubenswrapper[4715]: I1203 22:20:22.681316 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 22:20:22 crc kubenswrapper[4715]: I1203 22:20:22.691932 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 22:20:22 crc kubenswrapper[4715]: I1203 22:20:22.695658 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 22:20:24 crc kubenswrapper[4715]: I1203 22:20:24.634858 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:20:24 crc kubenswrapper[4715]: E1203 22:20:24.635314 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:20:32 crc kubenswrapper[4715]: I1203 22:20:32.794212 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:20:34 crc kubenswrapper[4715]: I1203 22:20:34.167647 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:20:36 crc kubenswrapper[4715]: I1203 22:20:36.932962 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerName="rabbitmq" containerID="cri-o://9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7" gracePeriod=604796 Dec 03 22:20:37 crc kubenswrapper[4715]: I1203 22:20:37.633985 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:20:37 crc kubenswrapper[4715]: E1203 22:20:37.634713 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:20:38 crc kubenswrapper[4715]: I1203 22:20:38.082393 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="d6205840-7ea5-4060-844a-f6ce9feed200" containerName="rabbitmq" containerID="cri-o://5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e" gracePeriod=604797 Dec 03 22:20:39 crc kubenswrapper[4715]: I1203 22:20:39.362274 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.96:5671: connect: connection refused" Dec 03 22:20:39 crc kubenswrapper[4715]: I1203 22:20:39.746669 4715 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="d6205840-7ea5-4060-844a-f6ce9feed200" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.97:5671: connect: connection refused" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.765786 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.775763 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837644 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-erlang-cookie-secret\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837697 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837727 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-plugins\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837768 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-confd\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837793 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-erlang-cookie\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837845 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmqlc\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-kube-api-access-cmqlc\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837876 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-server-conf\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837899 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-plugins\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837926 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-config-data\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.837990 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d6205840-7ea5-4060-844a-f6ce9feed200-pod-info\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838021 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-erlang-cookie\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838057 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-server-conf\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838090 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-config-data\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838157 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-confd\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838184 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-tls\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838215 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-894b7\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-kube-api-access-894b7\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838234 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838299 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-plugins-conf\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838365 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d6205840-7ea5-4060-844a-f6ce9feed200-erlang-cookie-secret\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838399 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-plugins-conf\") pod \"d6205840-7ea5-4060-844a-f6ce9feed200\" (UID: \"d6205840-7ea5-4060-844a-f6ce9feed200\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838427 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-tls\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.838454 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-pod-info\") pod \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\" (UID: \"5e2a0d28-27e4-4a86-8bf7-d5007af5162c\") " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.840736 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.845930 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.847139 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.848139 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.850717 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.850827 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-pod-info" (OuterVolumeSpecName: "pod-info") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.851427 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.856575 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.859580 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "persistence") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.860553 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d6205840-7ea5-4060-844a-f6ce9feed200-pod-info" (OuterVolumeSpecName: "pod-info") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.861419 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.866795 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.883492 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-kube-api-access-894b7" (OuterVolumeSpecName: "kube-api-access-894b7") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "kube-api-access-894b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.887960 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-kube-api-access-cmqlc" (OuterVolumeSpecName: "kube-api-access-cmqlc") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "kube-api-access-cmqlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.888096 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.889769 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6205840-7ea5-4060-844a-f6ce9feed200-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.908154 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-config-data" (OuterVolumeSpecName: "config-data") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.932572 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-config-data" (OuterVolumeSpecName: "config-data") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.945902 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-server-conf" (OuterVolumeSpecName: "server-conf") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.947060 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-server-conf" (OuterVolumeSpecName: "server-conf") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.949985 4715 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950019 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950031 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950070 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-894b7\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-kube-api-access-894b7\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950109 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950236 4715 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950250 4715 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d6205840-7ea5-4060-844a-f6ce9feed200-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950262 4715 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d6205840-7ea5-4060-844a-f6ce9feed200-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950273 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950530 4715 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950548 4715 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.950569 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.951544 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.951570 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.951639 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmqlc\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-kube-api-access-cmqlc\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.951657 4715 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.951669 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.951679 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.951691 4715 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d6205840-7ea5-4060-844a-f6ce9feed200-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.951705 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.965597 4715 generic.go:334] "Generic (PLEG): container finished" podID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerID="9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7" exitCode=0 Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.965883 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5e2a0d28-27e4-4a86-8bf7-d5007af5162c","Type":"ContainerDied","Data":"9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7"} Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.965978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"5e2a0d28-27e4-4a86-8bf7-d5007af5162c","Type":"ContainerDied","Data":"81d7bb66259f93ed78635e199b00099870547e3e598c3da06b3571c490a8b28c"} Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.966090 4715 scope.go:117] "RemoveContainer" containerID="9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.966984 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.977128 4715 generic.go:334] "Generic (PLEG): container finished" podID="d6205840-7ea5-4060-844a-f6ce9feed200" containerID="5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e" exitCode=0 Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.977168 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d6205840-7ea5-4060-844a-f6ce9feed200","Type":"ContainerDied","Data":"5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e"} Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.977192 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d6205840-7ea5-4060-844a-f6ce9feed200","Type":"ContainerDied","Data":"c2c285d63f37c200c967db8a0bdf6f3546f1f9cfa79a99c27918b3f77ad7aad9"} Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.977250 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.983398 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 03 22:20:45 crc kubenswrapper[4715]: I1203 22:20:45.996393 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.007088 4715 scope.go:117] "RemoveContainer" containerID="56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.030134 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "5e2a0d28-27e4-4a86-8bf7-d5007af5162c" (UID: "5e2a0d28-27e4-4a86-8bf7-d5007af5162c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.040902 4715 scope.go:117] "RemoveContainer" containerID="9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7" Dec 03 22:20:46 crc kubenswrapper[4715]: E1203 22:20:46.041376 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7\": container with ID starting with 9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7 not found: ID does not exist" containerID="9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.041418 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7"} err="failed to get container status \"9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7\": rpc error: code = NotFound desc = could not find container \"9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7\": container with ID starting with 9e82868f01d2bc0ff1df50b80f5a17f6abebac532209a2e46003e49aa182dda7 not found: ID does not exist" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.041445 4715 scope.go:117] "RemoveContainer" containerID="56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019" Dec 03 22:20:46 crc kubenswrapper[4715]: E1203 22:20:46.041885 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019\": container with ID starting with 56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019 not found: ID does not exist" containerID="56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.042027 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019"} err="failed to get container status \"56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019\": rpc error: code = NotFound desc = could not find container \"56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019\": container with ID starting with 56d6b4af0c1a046e8ed2a91345c96b050318bf8e8db1e66d6ace0bc9bbffd019 not found: ID does not exist" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.042134 4715 scope.go:117] "RemoveContainer" containerID="5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.053106 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.053278 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/5e2a0d28-27e4-4a86-8bf7-d5007af5162c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.053393 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.060288 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d6205840-7ea5-4060-844a-f6ce9feed200" (UID: "d6205840-7ea5-4060-844a-f6ce9feed200"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.067856 4715 scope.go:117] "RemoveContainer" containerID="f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.094826 4715 scope.go:117] "RemoveContainer" containerID="5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e" Dec 03 22:20:46 crc kubenswrapper[4715]: E1203 22:20:46.095238 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e\": container with ID starting with 5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e not found: ID does not exist" containerID="5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.095394 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e"} err="failed to get container status \"5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e\": rpc error: code = NotFound desc = could not find container \"5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e\": container with ID starting with 5275b1e86e8afc5e9ec8c4f8d8e71ca64879c7c5c2604f02c2c1267208abfd5e not found: ID does not exist" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.095428 4715 scope.go:117] "RemoveContainer" containerID="f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd" Dec 03 22:20:46 crc kubenswrapper[4715]: E1203 22:20:46.095802 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd\": container with ID starting with f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd not found: ID does not exist" containerID="f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.095833 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd"} err="failed to get container status \"f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd\": rpc error: code = NotFound desc = could not find container \"f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd\": container with ID starting with f0fa97d6a2907fbcc61d30899657f8998ef193fd5eb061a3c1edc36688f785bd not found: ID does not exist" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.155635 4715 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d6205840-7ea5-4060-844a-f6ce9feed200-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.303447 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.310888 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.319453 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.327831 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.336131 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:20:46 crc kubenswrapper[4715]: E1203 22:20:46.337115 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerName="rabbitmq" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.337260 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerName="rabbitmq" Dec 03 22:20:46 crc kubenswrapper[4715]: E1203 22:20:46.337399 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6205840-7ea5-4060-844a-f6ce9feed200" containerName="rabbitmq" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.337541 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6205840-7ea5-4060-844a-f6ce9feed200" containerName="rabbitmq" Dec 03 22:20:46 crc kubenswrapper[4715]: E1203 22:20:46.337780 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6205840-7ea5-4060-844a-f6ce9feed200" containerName="setup-container" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.337922 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6205840-7ea5-4060-844a-f6ce9feed200" containerName="setup-container" Dec 03 22:20:46 crc kubenswrapper[4715]: E1203 22:20:46.338095 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerName="setup-container" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.338210 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerName="setup-container" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.338679 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6205840-7ea5-4060-844a-f6ce9feed200" containerName="rabbitmq" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.338858 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" containerName="rabbitmq" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.340745 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.346055 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.346177 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-7pp49" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.346281 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.346487 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.347068 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.347301 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.347468 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.347666 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.348640 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.351864 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.352032 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.352239 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-8mhj2" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.352385 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.352391 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.352892 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.353162 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.353338 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.367092 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461023 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461074 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cb376763-8ff4-48ce-a374-ed1119c961e9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461109 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cb376763-8ff4-48ce-a374-ed1119c961e9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461229 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82483123-49b2-496c-96f4-2d9423e2dfeb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461323 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461360 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxh4s\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-kube-api-access-jxh4s\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461492 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-config-data\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461570 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82483123-49b2-496c-96f4-2d9423e2dfeb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461598 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461630 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461695 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461721 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wbgk\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-kube-api-access-9wbgk\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461835 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461868 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461911 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.461959 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.462005 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.462031 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.462068 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.462100 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.462122 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.462163 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564050 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-config-data\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564375 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82483123-49b2-496c-96f4-2d9423e2dfeb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564472 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564577 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564738 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wbgk\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-kube-api-access-9wbgk\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564910 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.564991 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565074 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565161 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565238 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565168 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565391 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565472 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565594 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565630 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565308 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565286 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565609 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565879 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565964 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.566044 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cb376763-8ff4-48ce-a374-ed1119c961e9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.566108 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cb376763-8ff4-48ce-a374-ed1119c961e9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.566184 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82483123-49b2-496c-96f4-2d9423e2dfeb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.566255 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.566265 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.565888 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.566345 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxh4s\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-kube-api-access-jxh4s\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.566720 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.566979 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.567278 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.567456 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cb376763-8ff4-48ce-a374-ed1119c961e9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.570264 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/82483123-49b2-496c-96f4-2d9423e2dfeb-config-data\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.571135 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/82483123-49b2-496c-96f4-2d9423e2dfeb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.572188 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.575124 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/82483123-49b2-496c-96f4-2d9423e2dfeb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.575934 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.576327 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.576674 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cb376763-8ff4-48ce-a374-ed1119c961e9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.576785 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cb376763-8ff4-48ce-a374-ed1119c961e9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.582203 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.585063 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxh4s\" (UniqueName: \"kubernetes.io/projected/cb376763-8ff4-48ce-a374-ed1119c961e9-kube-api-access-jxh4s\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.599910 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wbgk\" (UniqueName: \"kubernetes.io/projected/82483123-49b2-496c-96f4-2d9423e2dfeb-kube-api-access-9wbgk\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.613116 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cb376763-8ff4-48ce-a374-ed1119c961e9\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.617526 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"82483123-49b2-496c-96f4-2d9423e2dfeb\") " pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.663845 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 22:20:46 crc kubenswrapper[4715]: I1203 22:20:46.743818 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.003976 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.289890 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.410393 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-8wnf9"] Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.412128 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.414018 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.448995 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-8wnf9"] Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.495154 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.495213 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-config\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.495256 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.495337 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdtnt\" (UniqueName: \"kubernetes.io/projected/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-kube-api-access-wdtnt\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.495391 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.495446 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.495481 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-svc\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.596898 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-config\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.596950 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.597061 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdtnt\" (UniqueName: \"kubernetes.io/projected/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-kube-api-access-wdtnt\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.597105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.597154 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.597233 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-svc\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.597319 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.598229 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.598269 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-svc\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.598495 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.598597 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.599175 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-config\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.599200 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.614451 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdtnt\" (UniqueName: \"kubernetes.io/projected/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-kube-api-access-wdtnt\") pod \"dnsmasq-dns-d558885bc-8wnf9\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.646528 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2a0d28-27e4-4a86-8bf7-d5007af5162c" path="/var/lib/kubelet/pods/5e2a0d28-27e4-4a86-8bf7-d5007af5162c/volumes" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.647389 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6205840-7ea5-4060-844a-f6ce9feed200" path="/var/lib/kubelet/pods/d6205840-7ea5-4060-844a-f6ce9feed200/volumes" Dec 03 22:20:47 crc kubenswrapper[4715]: I1203 22:20:47.788550 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:48 crc kubenswrapper[4715]: I1203 22:20:48.023542 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cb376763-8ff4-48ce-a374-ed1119c961e9","Type":"ContainerStarted","Data":"8032b5bb76ba98b694a27245b3c415d49c99ab63c69684b4c42e79294bb6ed58"} Dec 03 22:20:48 crc kubenswrapper[4715]: I1203 22:20:48.026195 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82483123-49b2-496c-96f4-2d9423e2dfeb","Type":"ContainerStarted","Data":"1d494bd8179606c015e8264f1a967b7fe548e83bd3acc5f95446c809d3eeb7c5"} Dec 03 22:20:48 crc kubenswrapper[4715]: I1203 22:20:48.306868 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-8wnf9"] Dec 03 22:20:48 crc kubenswrapper[4715]: W1203 22:20:48.330926 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08ccb1fa_829f_47ca_bc15_2fc4e360e09e.slice/crio-ce109df7d01abbb4659826d997658d928ce3af6c39a1313c4ef8ac0041297afb WatchSource:0}: Error finding container ce109df7d01abbb4659826d997658d928ce3af6c39a1313c4ef8ac0041297afb: Status 404 returned error can't find the container with id ce109df7d01abbb4659826d997658d928ce3af6c39a1313c4ef8ac0041297afb Dec 03 22:20:49 crc kubenswrapper[4715]: I1203 22:20:49.045992 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82483123-49b2-496c-96f4-2d9423e2dfeb","Type":"ContainerStarted","Data":"98c1f139c73f1fe07607fee688b35e94ab14870f40b5a763cad9c8c2f501f85b"} Dec 03 22:20:49 crc kubenswrapper[4715]: I1203 22:20:49.049095 4715 generic.go:334] "Generic (PLEG): container finished" podID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" containerID="b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a" exitCode=0 Dec 03 22:20:49 crc kubenswrapper[4715]: I1203 22:20:49.049146 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" event={"ID":"08ccb1fa-829f-47ca-bc15-2fc4e360e09e","Type":"ContainerDied","Data":"b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a"} Dec 03 22:20:49 crc kubenswrapper[4715]: I1203 22:20:49.049698 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" event={"ID":"08ccb1fa-829f-47ca-bc15-2fc4e360e09e","Type":"ContainerStarted","Data":"ce109df7d01abbb4659826d997658d928ce3af6c39a1313c4ef8ac0041297afb"} Dec 03 22:20:49 crc kubenswrapper[4715]: I1203 22:20:49.635576 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:20:49 crc kubenswrapper[4715]: E1203 22:20:49.635884 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:20:50 crc kubenswrapper[4715]: I1203 22:20:50.064905 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" event={"ID":"08ccb1fa-829f-47ca-bc15-2fc4e360e09e","Type":"ContainerStarted","Data":"2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a"} Dec 03 22:20:50 crc kubenswrapper[4715]: I1203 22:20:50.065074 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:50 crc kubenswrapper[4715]: I1203 22:20:50.068184 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cb376763-8ff4-48ce-a374-ed1119c961e9","Type":"ContainerStarted","Data":"92c0bfadf1f5f2e923f83fdb1dbf3eae80233f041044cf4bb06a646ed6edb698"} Dec 03 22:20:50 crc kubenswrapper[4715]: I1203 22:20:50.091103 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" podStartSLOduration=3.091086854 podStartE2EDuration="3.091086854s" podCreationTimestamp="2025-12-03 22:20:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:20:50.089214924 +0000 UTC m=+2346.831925599" watchObservedRunningTime="2025-12-03 22:20:50.091086854 +0000 UTC m=+2346.833797459" Dec 03 22:20:57 crc kubenswrapper[4715]: I1203 22:20:57.789643 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:20:57 crc kubenswrapper[4715]: I1203 22:20:57.850104 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-fnbg8"] Dec 03 22:20:57 crc kubenswrapper[4715]: I1203 22:20:57.850462 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" podUID="08ba21f3-cfda-482a-b582-0cf75e86acea" containerName="dnsmasq-dns" containerID="cri-o://2d6390320ff45b454468fb873b15d44f83cd163844da43b681ba5423756ce064" gracePeriod=10 Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.087677 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-bhvj6"] Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.089430 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.134839 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-config\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.134955 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.134985 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.135036 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.135102 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.135130 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdxqv\" (UniqueName: \"kubernetes.io/projected/a1e5e31a-ff62-48c9-8908-15ff570bda0d-kube-api-access-mdxqv\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.135161 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.137620 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-bhvj6"] Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.182721 4715 generic.go:334] "Generic (PLEG): container finished" podID="08ba21f3-cfda-482a-b582-0cf75e86acea" containerID="2d6390320ff45b454468fb873b15d44f83cd163844da43b681ba5423756ce064" exitCode=0 Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.182819 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" event={"ID":"08ba21f3-cfda-482a-b582-0cf75e86acea","Type":"ContainerDied","Data":"2d6390320ff45b454468fb873b15d44f83cd163844da43b681ba5423756ce064"} Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.237960 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.238010 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.238054 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.238124 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.238148 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdxqv\" (UniqueName: \"kubernetes.io/projected/a1e5e31a-ff62-48c9-8908-15ff570bda0d-kube-api-access-mdxqv\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.238172 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.238215 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-config\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.239137 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-config\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.239693 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-openstack-edpm-ipam\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.240242 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-ovsdbserver-sb\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.240898 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-dns-svc\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.241420 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-ovsdbserver-nb\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.242107 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a1e5e31a-ff62-48c9-8908-15ff570bda0d-dns-swift-storage-0\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.266885 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdxqv\" (UniqueName: \"kubernetes.io/projected/a1e5e31a-ff62-48c9-8908-15ff570bda0d-kube-api-access-mdxqv\") pod \"dnsmasq-dns-78c64bc9c5-bhvj6\" (UID: \"a1e5e31a-ff62-48c9-8908-15ff570bda0d\") " pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.440398 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.555759 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.645095 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-sb\") pod \"08ba21f3-cfda-482a-b582-0cf75e86acea\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.645172 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb\") pod \"08ba21f3-cfda-482a-b582-0cf75e86acea\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.645202 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-config\") pod \"08ba21f3-cfda-482a-b582-0cf75e86acea\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.645251 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-swift-storage-0\") pod \"08ba21f3-cfda-482a-b582-0cf75e86acea\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.645335 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dbzm\" (UniqueName: \"kubernetes.io/projected/08ba21f3-cfda-482a-b582-0cf75e86acea-kube-api-access-9dbzm\") pod \"08ba21f3-cfda-482a-b582-0cf75e86acea\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.645401 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-svc\") pod \"08ba21f3-cfda-482a-b582-0cf75e86acea\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.650776 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08ba21f3-cfda-482a-b582-0cf75e86acea-kube-api-access-9dbzm" (OuterVolumeSpecName: "kube-api-access-9dbzm") pod "08ba21f3-cfda-482a-b582-0cf75e86acea" (UID: "08ba21f3-cfda-482a-b582-0cf75e86acea"). InnerVolumeSpecName "kube-api-access-9dbzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.695783 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08ba21f3-cfda-482a-b582-0cf75e86acea" (UID: "08ba21f3-cfda-482a-b582-0cf75e86acea"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.697551 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "08ba21f3-cfda-482a-b582-0cf75e86acea" (UID: "08ba21f3-cfda-482a-b582-0cf75e86acea"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.719566 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "08ba21f3-cfda-482a-b582-0cf75e86acea" (UID: "08ba21f3-cfda-482a-b582-0cf75e86acea"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:58 crc kubenswrapper[4715]: E1203 22:20:58.739824 4715 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb podName:08ba21f3-cfda-482a-b582-0cf75e86acea nodeName:}" failed. No retries permitted until 2025-12-03 22:20:59.239781632 +0000 UTC m=+2355.982492227 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-nb" (UniqueName: "kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb") pod "08ba21f3-cfda-482a-b582-0cf75e86acea" (UID: "08ba21f3-cfda-482a-b582-0cf75e86acea") : error deleting /var/lib/kubelet/pods/08ba21f3-cfda-482a-b582-0cf75e86acea/volume-subpaths: remove /var/lib/kubelet/pods/08ba21f3-cfda-482a-b582-0cf75e86acea/volume-subpaths: no such file or directory Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.740134 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-config" (OuterVolumeSpecName: "config") pod "08ba21f3-cfda-482a-b582-0cf75e86acea" (UID: "08ba21f3-cfda-482a-b582-0cf75e86acea"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.747914 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.747950 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.747962 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dbzm\" (UniqueName: \"kubernetes.io/projected/08ba21f3-cfda-482a-b582-0cf75e86acea-kube-api-access-9dbzm\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.747971 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.747979 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:58 crc kubenswrapper[4715]: I1203 22:20:58.888088 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78c64bc9c5-bhvj6"] Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.193465 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" event={"ID":"08ba21f3-cfda-482a-b582-0cf75e86acea","Type":"ContainerDied","Data":"16de09fd2273181d56a5983eb4929dbb0f1ab5dbef3ac925d77cc4f449653c4a"} Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.193781 4715 scope.go:117] "RemoveContainer" containerID="2d6390320ff45b454468fb873b15d44f83cd163844da43b681ba5423756ce064" Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.193536 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-fnbg8" Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.194978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" event={"ID":"a1e5e31a-ff62-48c9-8908-15ff570bda0d","Type":"ContainerStarted","Data":"74f27dca9ee38517e61046d0e0b04e2e487634f73482d3192fa67564cdcd6f44"} Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.213539 4715 scope.go:117] "RemoveContainer" containerID="94d628406d5f2b5b4249b4c5cff8bbb9a42ec6c461ad86713cf2a8412bcf2a89" Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.257568 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb\") pod \"08ba21f3-cfda-482a-b582-0cf75e86acea\" (UID: \"08ba21f3-cfda-482a-b582-0cf75e86acea\") " Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.258028 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "08ba21f3-cfda-482a-b582-0cf75e86acea" (UID: "08ba21f3-cfda-482a-b582-0cf75e86acea"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.258295 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ba21f3-cfda-482a-b582-0cf75e86acea-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.525958 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-fnbg8"] Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.535667 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-fnbg8"] Dec 03 22:20:59 crc kubenswrapper[4715]: I1203 22:20:59.645867 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08ba21f3-cfda-482a-b582-0cf75e86acea" path="/var/lib/kubelet/pods/08ba21f3-cfda-482a-b582-0cf75e86acea/volumes" Dec 03 22:21:00 crc kubenswrapper[4715]: I1203 22:21:00.210970 4715 generic.go:334] "Generic (PLEG): container finished" podID="a1e5e31a-ff62-48c9-8908-15ff570bda0d" containerID="eeb9b1c54636dd18f3c030aa63950ca5979774e6b376d943ba8d56ae969a8f8e" exitCode=0 Dec 03 22:21:00 crc kubenswrapper[4715]: I1203 22:21:00.211044 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" event={"ID":"a1e5e31a-ff62-48c9-8908-15ff570bda0d","Type":"ContainerDied","Data":"eeb9b1c54636dd18f3c030aa63950ca5979774e6b376d943ba8d56ae969a8f8e"} Dec 03 22:21:00 crc kubenswrapper[4715]: I1203 22:21:00.634915 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:21:00 crc kubenswrapper[4715]: E1203 22:21:00.635354 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:21:01 crc kubenswrapper[4715]: I1203 22:21:01.222828 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" event={"ID":"a1e5e31a-ff62-48c9-8908-15ff570bda0d","Type":"ContainerStarted","Data":"611f97c2fa69b264ae37c41386dd5822e715e0e8ed9ce82e5955578d77d9c331"} Dec 03 22:21:01 crc kubenswrapper[4715]: I1203 22:21:01.223293 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:21:01 crc kubenswrapper[4715]: I1203 22:21:01.245874 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" podStartSLOduration=3.245853991 podStartE2EDuration="3.245853991s" podCreationTimestamp="2025-12-03 22:20:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:21:01.243486148 +0000 UTC m=+2357.986196763" watchObservedRunningTime="2025-12-03 22:21:01.245853991 +0000 UTC m=+2357.988564586" Dec 03 22:21:08 crc kubenswrapper[4715]: I1203 22:21:08.443493 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-78c64bc9c5-bhvj6" Dec 03 22:21:08 crc kubenswrapper[4715]: I1203 22:21:08.525975 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-8wnf9"] Dec 03 22:21:08 crc kubenswrapper[4715]: I1203 22:21:08.526203 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" podUID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" containerName="dnsmasq-dns" containerID="cri-o://2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a" gracePeriod=10 Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.208229 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.318661 4715 generic.go:334] "Generic (PLEG): container finished" podID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" containerID="2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a" exitCode=0 Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.318700 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" event={"ID":"08ccb1fa-829f-47ca-bc15-2fc4e360e09e","Type":"ContainerDied","Data":"2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a"} Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.318724 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" event={"ID":"08ccb1fa-829f-47ca-bc15-2fc4e360e09e","Type":"ContainerDied","Data":"ce109df7d01abbb4659826d997658d928ce3af6c39a1313c4ef8ac0041297afb"} Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.318739 4715 scope.go:117] "RemoveContainer" containerID="2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.318847 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-8wnf9" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.341429 4715 scope.go:117] "RemoveContainer" containerID="b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.353957 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-sb\") pod \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.354907 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-swift-storage-0\") pod \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.355135 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-config\") pod \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.355245 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-svc\") pod \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.355581 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdtnt\" (UniqueName: \"kubernetes.io/projected/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-kube-api-access-wdtnt\") pod \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.355790 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-openstack-edpm-ipam\") pod \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.355921 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-nb\") pod \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\" (UID: \"08ccb1fa-829f-47ca-bc15-2fc4e360e09e\") " Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.360176 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-kube-api-access-wdtnt" (OuterVolumeSpecName: "kube-api-access-wdtnt") pod "08ccb1fa-829f-47ca-bc15-2fc4e360e09e" (UID: "08ccb1fa-829f-47ca-bc15-2fc4e360e09e"). InnerVolumeSpecName "kube-api-access-wdtnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.373227 4715 scope.go:117] "RemoveContainer" containerID="2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a" Dec 03 22:21:10 crc kubenswrapper[4715]: E1203 22:21:10.374198 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a\": container with ID starting with 2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a not found: ID does not exist" containerID="2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.374232 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a"} err="failed to get container status \"2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a\": rpc error: code = NotFound desc = could not find container \"2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a\": container with ID starting with 2371502a8441432ef485e22d1292e6dd07eac96a837f5abfd2fcc1814f51f07a not found: ID does not exist" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.374253 4715 scope.go:117] "RemoveContainer" containerID="b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a" Dec 03 22:21:10 crc kubenswrapper[4715]: E1203 22:21:10.375627 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a\": container with ID starting with b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a not found: ID does not exist" containerID="b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.375688 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a"} err="failed to get container status \"b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a\": rpc error: code = NotFound desc = could not find container \"b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a\": container with ID starting with b7188f7b2839fe6a5c5a126c0a1357493183ba22cb67f8482b6619b082bc843a not found: ID does not exist" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.408895 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "08ccb1fa-829f-47ca-bc15-2fc4e360e09e" (UID: "08ccb1fa-829f-47ca-bc15-2fc4e360e09e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.415518 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "08ccb1fa-829f-47ca-bc15-2fc4e360e09e" (UID: "08ccb1fa-829f-47ca-bc15-2fc4e360e09e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.415739 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "08ccb1fa-829f-47ca-bc15-2fc4e360e09e" (UID: "08ccb1fa-829f-47ca-bc15-2fc4e360e09e"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.430912 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-config" (OuterVolumeSpecName: "config") pod "08ccb1fa-829f-47ca-bc15-2fc4e360e09e" (UID: "08ccb1fa-829f-47ca-bc15-2fc4e360e09e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.436098 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "08ccb1fa-829f-47ca-bc15-2fc4e360e09e" (UID: "08ccb1fa-829f-47ca-bc15-2fc4e360e09e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.437978 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "08ccb1fa-829f-47ca-bc15-2fc4e360e09e" (UID: "08ccb1fa-829f-47ca-bc15-2fc4e360e09e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.460908 4715 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.460936 4715 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.460945 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdtnt\" (UniqueName: \"kubernetes.io/projected/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-kube-api-access-wdtnt\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.460955 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.460964 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.460972 4715 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.460979 4715 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/08ccb1fa-829f-47ca-bc15-2fc4e360e09e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.672221 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-8wnf9"] Dec 03 22:21:10 crc kubenswrapper[4715]: I1203 22:21:10.685769 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-8wnf9"] Dec 03 22:21:11 crc kubenswrapper[4715]: I1203 22:21:11.649015 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" path="/var/lib/kubelet/pods/08ccb1fa-829f-47ca-bc15-2fc4e360e09e/volumes" Dec 03 22:21:12 crc kubenswrapper[4715]: I1203 22:21:12.634520 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:21:12 crc kubenswrapper[4715]: E1203 22:21:12.635145 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.733133 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9"] Dec 03 22:21:20 crc kubenswrapper[4715]: E1203 22:21:20.734242 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ba21f3-cfda-482a-b582-0cf75e86acea" containerName="dnsmasq-dns" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.734441 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ba21f3-cfda-482a-b582-0cf75e86acea" containerName="dnsmasq-dns" Dec 03 22:21:20 crc kubenswrapper[4715]: E1203 22:21:20.734460 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" containerName="dnsmasq-dns" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.734469 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" containerName="dnsmasq-dns" Dec 03 22:21:20 crc kubenswrapper[4715]: E1203 22:21:20.734527 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ba21f3-cfda-482a-b582-0cf75e86acea" containerName="init" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.734540 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ba21f3-cfda-482a-b582-0cf75e86acea" containerName="init" Dec 03 22:21:20 crc kubenswrapper[4715]: E1203 22:21:20.734582 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" containerName="init" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.734593 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" containerName="init" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.734889 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ccb1fa-829f-47ca-bc15-2fc4e360e09e" containerName="dnsmasq-dns" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.734940 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ba21f3-cfda-482a-b582-0cf75e86acea" containerName="dnsmasq-dns" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.735888 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.738865 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.739338 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.739635 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.741307 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.760621 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9"] Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.808355 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.808445 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.808483 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvgj2\" (UniqueName: \"kubernetes.io/projected/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-kube-api-access-kvgj2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.808580 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.910197 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.910580 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.910622 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvgj2\" (UniqueName: \"kubernetes.io/projected/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-kube-api-access-kvgj2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.910715 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.916951 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.917853 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.921033 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:20 crc kubenswrapper[4715]: I1203 22:21:20.925675 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvgj2\" (UniqueName: \"kubernetes.io/projected/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-kube-api-access-kvgj2\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:21 crc kubenswrapper[4715]: I1203 22:21:21.081528 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:21 crc kubenswrapper[4715]: I1203 22:21:21.680176 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9"] Dec 03 22:21:21 crc kubenswrapper[4715]: W1203 22:21:21.688950 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a15fc43_6ab8_4502_a2d3_9c837a74cb90.slice/crio-0de0ee7d068a19bcef6245c63b5cd4116d6124f475f0645a1e5f98b9fdfa5f84 WatchSource:0}: Error finding container 0de0ee7d068a19bcef6245c63b5cd4116d6124f475f0645a1e5f98b9fdfa5f84: Status 404 returned error can't find the container with id 0de0ee7d068a19bcef6245c63b5cd4116d6124f475f0645a1e5f98b9fdfa5f84 Dec 03 22:21:22 crc kubenswrapper[4715]: I1203 22:21:22.459639 4715 generic.go:334] "Generic (PLEG): container finished" podID="82483123-49b2-496c-96f4-2d9423e2dfeb" containerID="98c1f139c73f1fe07607fee688b35e94ab14870f40b5a763cad9c8c2f501f85b" exitCode=0 Dec 03 22:21:22 crc kubenswrapper[4715]: I1203 22:21:22.459827 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82483123-49b2-496c-96f4-2d9423e2dfeb","Type":"ContainerDied","Data":"98c1f139c73f1fe07607fee688b35e94ab14870f40b5a763cad9c8c2f501f85b"} Dec 03 22:21:22 crc kubenswrapper[4715]: I1203 22:21:22.482551 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" event={"ID":"4a15fc43-6ab8-4502-a2d3-9c837a74cb90","Type":"ContainerStarted","Data":"0de0ee7d068a19bcef6245c63b5cd4116d6124f475f0645a1e5f98b9fdfa5f84"} Dec 03 22:21:23 crc kubenswrapper[4715]: I1203 22:21:23.496042 4715 generic.go:334] "Generic (PLEG): container finished" podID="cb376763-8ff4-48ce-a374-ed1119c961e9" containerID="92c0bfadf1f5f2e923f83fdb1dbf3eae80233f041044cf4bb06a646ed6edb698" exitCode=0 Dec 03 22:21:23 crc kubenswrapper[4715]: I1203 22:21:23.496173 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cb376763-8ff4-48ce-a374-ed1119c961e9","Type":"ContainerDied","Data":"92c0bfadf1f5f2e923f83fdb1dbf3eae80233f041044cf4bb06a646ed6edb698"} Dec 03 22:21:25 crc kubenswrapper[4715]: I1203 22:21:24.509279 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"82483123-49b2-496c-96f4-2d9423e2dfeb","Type":"ContainerStarted","Data":"9353ba8b45860e71317378f6f9293a14542efa2fe41f9329f076d0dcac8ee61c"} Dec 03 22:21:25 crc kubenswrapper[4715]: I1203 22:21:25.520929 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cb376763-8ff4-48ce-a374-ed1119c961e9","Type":"ContainerStarted","Data":"37f99c5310f3a8402403b8a43bddab9390746abe8dc1f035d594c5ab76086c5a"} Dec 03 22:21:25 crc kubenswrapper[4715]: I1203 22:21:25.521245 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 22:21:25 crc kubenswrapper[4715]: I1203 22:21:25.521426 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:21:25 crc kubenswrapper[4715]: I1203 22:21:25.546390 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.546348349 podStartE2EDuration="39.546348349s" podCreationTimestamp="2025-12-03 22:20:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:21:25.540468162 +0000 UTC m=+2382.283178767" watchObservedRunningTime="2025-12-03 22:21:25.546348349 +0000 UTC m=+2382.289058934" Dec 03 22:21:25 crc kubenswrapper[4715]: I1203 22:21:25.580718 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.580701138 podStartE2EDuration="39.580701138s" podCreationTimestamp="2025-12-03 22:20:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 22:21:25.572303574 +0000 UTC m=+2382.315014169" watchObservedRunningTime="2025-12-03 22:21:25.580701138 +0000 UTC m=+2382.323411723" Dec 03 22:21:25 crc kubenswrapper[4715]: I1203 22:21:25.634666 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:21:25 crc kubenswrapper[4715]: E1203 22:21:25.634909 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:21:33 crc kubenswrapper[4715]: I1203 22:21:33.613661 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" event={"ID":"4a15fc43-6ab8-4502-a2d3-9c837a74cb90","Type":"ContainerStarted","Data":"3d6e655319c2fc3cb1b8db7cee8e249ee92bbd9f9563d678612550e3b11c825e"} Dec 03 22:21:33 crc kubenswrapper[4715]: I1203 22:21:33.650893 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" podStartSLOduration=2.380220409 podStartE2EDuration="13.650873305s" podCreationTimestamp="2025-12-03 22:21:20 +0000 UTC" firstStartedPulling="2025-12-03 22:21:21.691995184 +0000 UTC m=+2378.434705779" lastFinishedPulling="2025-12-03 22:21:32.96264808 +0000 UTC m=+2389.705358675" observedRunningTime="2025-12-03 22:21:33.63312193 +0000 UTC m=+2390.375832525" watchObservedRunningTime="2025-12-03 22:21:33.650873305 +0000 UTC m=+2390.393583910" Dec 03 22:21:36 crc kubenswrapper[4715]: I1203 22:21:36.669678 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 22:21:36 crc kubenswrapper[4715]: I1203 22:21:36.749982 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 22:21:39 crc kubenswrapper[4715]: I1203 22:21:39.635552 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:21:39 crc kubenswrapper[4715]: E1203 22:21:39.636718 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:21:45 crc kubenswrapper[4715]: I1203 22:21:45.745731 4715 generic.go:334] "Generic (PLEG): container finished" podID="4a15fc43-6ab8-4502-a2d3-9c837a74cb90" containerID="3d6e655319c2fc3cb1b8db7cee8e249ee92bbd9f9563d678612550e3b11c825e" exitCode=0 Dec 03 22:21:45 crc kubenswrapper[4715]: I1203 22:21:45.745845 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" event={"ID":"4a15fc43-6ab8-4502-a2d3-9c837a74cb90","Type":"ContainerDied","Data":"3d6e655319c2fc3cb1b8db7cee8e249ee92bbd9f9563d678612550e3b11c825e"} Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.347645 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.445324 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-inventory\") pod \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.445666 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-ssh-key\") pod \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.445714 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-repo-setup-combined-ca-bundle\") pod \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.445763 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvgj2\" (UniqueName: \"kubernetes.io/projected/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-kube-api-access-kvgj2\") pod \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\" (UID: \"4a15fc43-6ab8-4502-a2d3-9c837a74cb90\") " Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.452430 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4a15fc43-6ab8-4502-a2d3-9c837a74cb90" (UID: "4a15fc43-6ab8-4502-a2d3-9c837a74cb90"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.453375 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-kube-api-access-kvgj2" (OuterVolumeSpecName: "kube-api-access-kvgj2") pod "4a15fc43-6ab8-4502-a2d3-9c837a74cb90" (UID: "4a15fc43-6ab8-4502-a2d3-9c837a74cb90"). InnerVolumeSpecName "kube-api-access-kvgj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.486556 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-inventory" (OuterVolumeSpecName: "inventory") pod "4a15fc43-6ab8-4502-a2d3-9c837a74cb90" (UID: "4a15fc43-6ab8-4502-a2d3-9c837a74cb90"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.486964 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a15fc43-6ab8-4502-a2d3-9c837a74cb90" (UID: "4a15fc43-6ab8-4502-a2d3-9c837a74cb90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.548923 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.549571 4715 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.549597 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvgj2\" (UniqueName: \"kubernetes.io/projected/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-kube-api-access-kvgj2\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.549659 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4a15fc43-6ab8-4502-a2d3-9c837a74cb90-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.772340 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" event={"ID":"4a15fc43-6ab8-4502-a2d3-9c837a74cb90","Type":"ContainerDied","Data":"0de0ee7d068a19bcef6245c63b5cd4116d6124f475f0645a1e5f98b9fdfa5f84"} Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.772385 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0de0ee7d068a19bcef6245c63b5cd4116d6124f475f0645a1e5f98b9fdfa5f84" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.772431 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.892419 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms"] Dec 03 22:21:47 crc kubenswrapper[4715]: E1203 22:21:47.893136 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a15fc43-6ab8-4502-a2d3-9c837a74cb90" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.893171 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a15fc43-6ab8-4502-a2d3-9c837a74cb90" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.893610 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a15fc43-6ab8-4502-a2d3-9c837a74cb90" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.894633 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.896884 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.897017 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.897389 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.898600 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.911936 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms"] Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.962234 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.962284 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:47 crc kubenswrapper[4715]: I1203 22:21:47.962527 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59vfk\" (UniqueName: \"kubernetes.io/projected/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-kube-api-access-59vfk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.064740 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.064819 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.065034 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59vfk\" (UniqueName: \"kubernetes.io/projected/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-kube-api-access-59vfk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.070290 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.070866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.089156 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59vfk\" (UniqueName: \"kubernetes.io/projected/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-kube-api-access-59vfk\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-8d7ms\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.223226 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.826946 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:21:48 crc kubenswrapper[4715]: I1203 22:21:48.830492 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms"] Dec 03 22:21:49 crc kubenswrapper[4715]: I1203 22:21:49.798119 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" event={"ID":"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93","Type":"ContainerStarted","Data":"063bb94d2b42273f9d1054a3fcb3e5df2f159ccd4f8d26efe5a21bd633ca38b3"} Dec 03 22:21:51 crc kubenswrapper[4715]: I1203 22:21:51.838578 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" event={"ID":"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93","Type":"ContainerStarted","Data":"780a58e2640e65874981c4560e46c12d004f5471be218cdda7c3545795809f42"} Dec 03 22:21:51 crc kubenswrapper[4715]: I1203 22:21:51.863193 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" podStartSLOduration=2.755548716 podStartE2EDuration="4.863171991s" podCreationTimestamp="2025-12-03 22:21:47 +0000 UTC" firstStartedPulling="2025-12-03 22:21:48.826570342 +0000 UTC m=+2405.569280977" lastFinishedPulling="2025-12-03 22:21:50.934193657 +0000 UTC m=+2407.676904252" observedRunningTime="2025-12-03 22:21:51.860427847 +0000 UTC m=+2408.603138472" watchObservedRunningTime="2025-12-03 22:21:51.863171991 +0000 UTC m=+2408.605882626" Dec 03 22:21:53 crc kubenswrapper[4715]: I1203 22:21:53.211406 4715 scope.go:117] "RemoveContainer" containerID="b98f9d2ae2e384b9468509f35809ef612a35199ef9b1d08ad4f67bb0812d635f" Dec 03 22:21:53 crc kubenswrapper[4715]: I1203 22:21:53.635827 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:21:53 crc kubenswrapper[4715]: E1203 22:21:53.636318 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:21:54 crc kubenswrapper[4715]: I1203 22:21:54.899220 4715 generic.go:334] "Generic (PLEG): container finished" podID="e1c0be7c-5e1c-43dd-9148-30cd6bc42b93" containerID="780a58e2640e65874981c4560e46c12d004f5471be218cdda7c3545795809f42" exitCode=0 Dec 03 22:21:54 crc kubenswrapper[4715]: I1203 22:21:54.899739 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" event={"ID":"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93","Type":"ContainerDied","Data":"780a58e2640e65874981c4560e46c12d004f5471be218cdda7c3545795809f42"} Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.411441 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.445295 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59vfk\" (UniqueName: \"kubernetes.io/projected/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-kube-api-access-59vfk\") pod \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.445419 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-inventory\") pod \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.445573 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-ssh-key\") pod \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\" (UID: \"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93\") " Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.454261 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-kube-api-access-59vfk" (OuterVolumeSpecName: "kube-api-access-59vfk") pod "e1c0be7c-5e1c-43dd-9148-30cd6bc42b93" (UID: "e1c0be7c-5e1c-43dd-9148-30cd6bc42b93"). InnerVolumeSpecName "kube-api-access-59vfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.492877 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-inventory" (OuterVolumeSpecName: "inventory") pod "e1c0be7c-5e1c-43dd-9148-30cd6bc42b93" (UID: "e1c0be7c-5e1c-43dd-9148-30cd6bc42b93"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.496637 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e1c0be7c-5e1c-43dd-9148-30cd6bc42b93" (UID: "e1c0be7c-5e1c-43dd-9148-30cd6bc42b93"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.548478 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59vfk\" (UniqueName: \"kubernetes.io/projected/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-kube-api-access-59vfk\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.548530 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.548542 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e1c0be7c-5e1c-43dd-9148-30cd6bc42b93-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.921664 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" event={"ID":"e1c0be7c-5e1c-43dd-9148-30cd6bc42b93","Type":"ContainerDied","Data":"063bb94d2b42273f9d1054a3fcb3e5df2f159ccd4f8d26efe5a21bd633ca38b3"} Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.922037 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="063bb94d2b42273f9d1054a3fcb3e5df2f159ccd4f8d26efe5a21bd633ca38b3" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:56.921706 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-8d7ms" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.010039 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx"] Dec 03 22:21:57 crc kubenswrapper[4715]: E1203 22:21:57.010484 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1c0be7c-5e1c-43dd-9148-30cd6bc42b93" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.010495 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1c0be7c-5e1c-43dd-9148-30cd6bc42b93" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.010707 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1c0be7c-5e1c-43dd-9148-30cd6bc42b93" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.011383 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.013253 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.013519 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.013495 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.013641 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.022589 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx"] Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.060124 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.060188 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlk5z\" (UniqueName: \"kubernetes.io/projected/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-kube-api-access-zlk5z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.060323 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.060404 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.162673 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.162753 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.162799 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlk5z\" (UniqueName: \"kubernetes.io/projected/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-kube-api-access-zlk5z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.162853 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.166878 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.167839 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.169181 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.180869 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlk5z\" (UniqueName: \"kubernetes.io/projected/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-kube-api-access-zlk5z\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.329264 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.908126 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx"] Dec 03 22:21:57 crc kubenswrapper[4715]: W1203 22:21:57.917441 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d5be5a7_46b7_4c15_b6ef_09e00d9a5b63.slice/crio-de01b7c929faa1e76be901f6ff51a977f8779a66e4e806b38fe615bb100fcf0f WatchSource:0}: Error finding container de01b7c929faa1e76be901f6ff51a977f8779a66e4e806b38fe615bb100fcf0f: Status 404 returned error can't find the container with id de01b7c929faa1e76be901f6ff51a977f8779a66e4e806b38fe615bb100fcf0f Dec 03 22:21:57 crc kubenswrapper[4715]: I1203 22:21:57.930320 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" event={"ID":"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63","Type":"ContainerStarted","Data":"de01b7c929faa1e76be901f6ff51a977f8779a66e4e806b38fe615bb100fcf0f"} Dec 03 22:22:01 crc kubenswrapper[4715]: I1203 22:22:01.972436 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" event={"ID":"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63","Type":"ContainerStarted","Data":"e9dd62558b17ee3cffb973e0885cfe2a0fce5de816f694f2fa8b834ecd04b851"} Dec 03 22:22:02 crc kubenswrapper[4715]: I1203 22:22:02.005245 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" podStartSLOduration=3.227293918 podStartE2EDuration="6.005212077s" podCreationTimestamp="2025-12-03 22:21:56 +0000 UTC" firstStartedPulling="2025-12-03 22:21:57.921496058 +0000 UTC m=+2414.664206653" lastFinishedPulling="2025-12-03 22:22:00.699414217 +0000 UTC m=+2417.442124812" observedRunningTime="2025-12-03 22:22:01.990595796 +0000 UTC m=+2418.733306451" watchObservedRunningTime="2025-12-03 22:22:02.005212077 +0000 UTC m=+2418.747922742" Dec 03 22:22:06 crc kubenswrapper[4715]: I1203 22:22:06.635283 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:22:06 crc kubenswrapper[4715]: E1203 22:22:06.636415 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:22:18 crc kubenswrapper[4715]: I1203 22:22:18.634483 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:22:18 crc kubenswrapper[4715]: E1203 22:22:18.635438 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:22:31 crc kubenswrapper[4715]: I1203 22:22:31.635318 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:22:31 crc kubenswrapper[4715]: E1203 22:22:31.636278 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:22:43 crc kubenswrapper[4715]: I1203 22:22:43.650654 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:22:43 crc kubenswrapper[4715]: E1203 22:22:43.651841 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:22:53 crc kubenswrapper[4715]: I1203 22:22:53.339878 4715 scope.go:117] "RemoveContainer" containerID="71de947a8dec65c3ee92f5feadf901bb7d74f27f725fdfb0d808ce3a07822f50" Dec 03 22:22:53 crc kubenswrapper[4715]: I1203 22:22:53.390212 4715 scope.go:117] "RemoveContainer" containerID="a5b0d903b1f711b99655d67b5066c14cddb95948e7c959d7124c9576077583ba" Dec 03 22:22:53 crc kubenswrapper[4715]: I1203 22:22:53.437159 4715 scope.go:117] "RemoveContainer" containerID="aed158e03dfae84d343df09491228128b814863fd73cf6b514deb3ccef7a9767" Dec 03 22:22:57 crc kubenswrapper[4715]: I1203 22:22:57.635828 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:22:57 crc kubenswrapper[4715]: E1203 22:22:57.636944 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:23:12 crc kubenswrapper[4715]: I1203 22:23:12.635566 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:23:12 crc kubenswrapper[4715]: E1203 22:23:12.636827 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:23:25 crc kubenswrapper[4715]: I1203 22:23:25.634014 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:23:25 crc kubenswrapper[4715]: E1203 22:23:25.634634 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:23:38 crc kubenswrapper[4715]: I1203 22:23:38.578653 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rlg4f"] Dec 03 22:23:38 crc kubenswrapper[4715]: I1203 22:23:38.581485 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:38 crc kubenswrapper[4715]: I1203 22:23:38.612048 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rlg4f"] Dec 03 22:23:38 crc kubenswrapper[4715]: I1203 22:23:38.635008 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:23:38 crc kubenswrapper[4715]: E1203 22:23:38.635544 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:23:38 crc kubenswrapper[4715]: I1203 22:23:38.657262 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-catalog-content\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:38 crc kubenswrapper[4715]: I1203 22:23:38.657587 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-utilities\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:38 crc kubenswrapper[4715]: I1203 22:23:38.657718 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-674j9\" (UniqueName: \"kubernetes.io/projected/2bfcec28-689c-4778-b9a6-2534ffb368b7-kube-api-access-674j9\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.064880 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-utilities\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.065826 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-674j9\" (UniqueName: \"kubernetes.io/projected/2bfcec28-689c-4778-b9a6-2534ffb368b7-kube-api-access-674j9\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.065866 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-utilities\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.066355 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-catalog-content\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.066888 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-catalog-content\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.103289 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-674j9\" (UniqueName: \"kubernetes.io/projected/2bfcec28-689c-4778-b9a6-2534ffb368b7-kube-api-access-674j9\") pod \"community-operators-rlg4f\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.217944 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.383919 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tpr6h"] Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.396676 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.413963 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpr6h"] Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.484476 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-catalog-content\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.484825 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-utilities\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.485064 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqwlf\" (UniqueName: \"kubernetes.io/projected/0ac36e60-4df8-4a07-8341-a77eb105e998-kube-api-access-lqwlf\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.586751 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-catalog-content\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.586848 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-utilities\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.586899 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqwlf\" (UniqueName: \"kubernetes.io/projected/0ac36e60-4df8-4a07-8341-a77eb105e998-kube-api-access-lqwlf\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.587827 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-catalog-content\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.588055 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-utilities\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.612207 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqwlf\" (UniqueName: \"kubernetes.io/projected/0ac36e60-4df8-4a07-8341-a77eb105e998-kube-api-access-lqwlf\") pod \"redhat-marketplace-tpr6h\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.733676 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:39 crc kubenswrapper[4715]: I1203 22:23:39.751610 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rlg4f"] Dec 03 22:23:40 crc kubenswrapper[4715]: I1203 22:23:40.170126 4715 generic.go:334] "Generic (PLEG): container finished" podID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerID="db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f" exitCode=0 Dec 03 22:23:40 crc kubenswrapper[4715]: I1203 22:23:40.170214 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlg4f" event={"ID":"2bfcec28-689c-4778-b9a6-2534ffb368b7","Type":"ContainerDied","Data":"db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f"} Dec 03 22:23:40 crc kubenswrapper[4715]: I1203 22:23:40.170411 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlg4f" event={"ID":"2bfcec28-689c-4778-b9a6-2534ffb368b7","Type":"ContainerStarted","Data":"3894c5b8ff1b903f962f93baa4ac4740f6620b3e4a9024be3d03da18e0ee5c97"} Dec 03 22:23:40 crc kubenswrapper[4715]: I1203 22:23:40.216302 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpr6h"] Dec 03 22:23:41 crc kubenswrapper[4715]: I1203 22:23:41.185150 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlg4f" event={"ID":"2bfcec28-689c-4778-b9a6-2534ffb368b7","Type":"ContainerStarted","Data":"fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe"} Dec 03 22:23:41 crc kubenswrapper[4715]: I1203 22:23:41.187365 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerID="c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d" exitCode=0 Dec 03 22:23:41 crc kubenswrapper[4715]: I1203 22:23:41.187424 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpr6h" event={"ID":"0ac36e60-4df8-4a07-8341-a77eb105e998","Type":"ContainerDied","Data":"c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d"} Dec 03 22:23:41 crc kubenswrapper[4715]: I1203 22:23:41.187462 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpr6h" event={"ID":"0ac36e60-4df8-4a07-8341-a77eb105e998","Type":"ContainerStarted","Data":"3da4a11b787231d9756c12d69714020e601df3bfa49d1c2017254329d69a777b"} Dec 03 22:23:42 crc kubenswrapper[4715]: I1203 22:23:42.199854 4715 generic.go:334] "Generic (PLEG): container finished" podID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerID="fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe" exitCode=0 Dec 03 22:23:42 crc kubenswrapper[4715]: I1203 22:23:42.199973 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlg4f" event={"ID":"2bfcec28-689c-4778-b9a6-2534ffb368b7","Type":"ContainerDied","Data":"fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe"} Dec 03 22:23:43 crc kubenswrapper[4715]: I1203 22:23:43.213919 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlg4f" event={"ID":"2bfcec28-689c-4778-b9a6-2534ffb368b7","Type":"ContainerStarted","Data":"d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb"} Dec 03 22:23:43 crc kubenswrapper[4715]: I1203 22:23:43.215985 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerID="30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53" exitCode=0 Dec 03 22:23:43 crc kubenswrapper[4715]: I1203 22:23:43.216035 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpr6h" event={"ID":"0ac36e60-4df8-4a07-8341-a77eb105e998","Type":"ContainerDied","Data":"30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53"} Dec 03 22:23:43 crc kubenswrapper[4715]: I1203 22:23:43.235368 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rlg4f" podStartSLOduration=2.761918009 podStartE2EDuration="5.235352655s" podCreationTimestamp="2025-12-03 22:23:38 +0000 UTC" firstStartedPulling="2025-12-03 22:23:40.171757667 +0000 UTC m=+2516.914468262" lastFinishedPulling="2025-12-03 22:23:42.645192313 +0000 UTC m=+2519.387902908" observedRunningTime="2025-12-03 22:23:43.232299394 +0000 UTC m=+2519.975009989" watchObservedRunningTime="2025-12-03 22:23:43.235352655 +0000 UTC m=+2519.978063250" Dec 03 22:23:44 crc kubenswrapper[4715]: I1203 22:23:44.225879 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpr6h" event={"ID":"0ac36e60-4df8-4a07-8341-a77eb105e998","Type":"ContainerStarted","Data":"74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56"} Dec 03 22:23:44 crc kubenswrapper[4715]: I1203 22:23:44.242544 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tpr6h" podStartSLOduration=2.764268425 podStartE2EDuration="5.24252125s" podCreationTimestamp="2025-12-03 22:23:39 +0000 UTC" firstStartedPulling="2025-12-03 22:23:41.190132411 +0000 UTC m=+2517.932843006" lastFinishedPulling="2025-12-03 22:23:43.668385226 +0000 UTC m=+2520.411095831" observedRunningTime="2025-12-03 22:23:44.241635076 +0000 UTC m=+2520.984345681" watchObservedRunningTime="2025-12-03 22:23:44.24252125 +0000 UTC m=+2520.985231855" Dec 03 22:23:49 crc kubenswrapper[4715]: I1203 22:23:49.218676 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:49 crc kubenswrapper[4715]: I1203 22:23:49.219251 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:49 crc kubenswrapper[4715]: I1203 22:23:49.279421 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:49 crc kubenswrapper[4715]: I1203 22:23:49.326056 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:49 crc kubenswrapper[4715]: I1203 22:23:49.516051 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rlg4f"] Dec 03 22:23:49 crc kubenswrapper[4715]: I1203 22:23:49.734401 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:49 crc kubenswrapper[4715]: I1203 22:23:49.734804 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:49 crc kubenswrapper[4715]: I1203 22:23:49.850790 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:50 crc kubenswrapper[4715]: I1203 22:23:50.446858 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:50 crc kubenswrapper[4715]: I1203 22:23:50.634624 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:23:50 crc kubenswrapper[4715]: E1203 22:23:50.634907 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:23:51 crc kubenswrapper[4715]: I1203 22:23:51.300606 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rlg4f" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerName="registry-server" containerID="cri-o://d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb" gracePeriod=2 Dec 03 22:23:52 crc kubenswrapper[4715]: I1203 22:23:52.723227 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpr6h"] Dec 03 22:23:52 crc kubenswrapper[4715]: I1203 22:23:52.785597 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:52 crc kubenswrapper[4715]: I1203 22:23:52.946905 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-catalog-content\") pod \"2bfcec28-689c-4778-b9a6-2534ffb368b7\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " Dec 03 22:23:52 crc kubenswrapper[4715]: I1203 22:23:52.947096 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-674j9\" (UniqueName: \"kubernetes.io/projected/2bfcec28-689c-4778-b9a6-2534ffb368b7-kube-api-access-674j9\") pod \"2bfcec28-689c-4778-b9a6-2534ffb368b7\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " Dec 03 22:23:52 crc kubenswrapper[4715]: I1203 22:23:52.947127 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-utilities\") pod \"2bfcec28-689c-4778-b9a6-2534ffb368b7\" (UID: \"2bfcec28-689c-4778-b9a6-2534ffb368b7\") " Dec 03 22:23:52 crc kubenswrapper[4715]: I1203 22:23:52.948188 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-utilities" (OuterVolumeSpecName: "utilities") pod "2bfcec28-689c-4778-b9a6-2534ffb368b7" (UID: "2bfcec28-689c-4778-b9a6-2534ffb368b7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:23:52 crc kubenswrapper[4715]: I1203 22:23:52.955130 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bfcec28-689c-4778-b9a6-2534ffb368b7-kube-api-access-674j9" (OuterVolumeSpecName: "kube-api-access-674j9") pod "2bfcec28-689c-4778-b9a6-2534ffb368b7" (UID: "2bfcec28-689c-4778-b9a6-2534ffb368b7"). InnerVolumeSpecName "kube-api-access-674j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.003427 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2bfcec28-689c-4778-b9a6-2534ffb368b7" (UID: "2bfcec28-689c-4778-b9a6-2534ffb368b7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.049687 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.049760 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-674j9\" (UniqueName: \"kubernetes.io/projected/2bfcec28-689c-4778-b9a6-2534ffb368b7-kube-api-access-674j9\") on node \"crc\" DevicePath \"\"" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.049789 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2bfcec28-689c-4778-b9a6-2534ffb368b7-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.322541 4715 generic.go:334] "Generic (PLEG): container finished" podID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerID="d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb" exitCode=0 Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.322682 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlg4f" event={"ID":"2bfcec28-689c-4778-b9a6-2534ffb368b7","Type":"ContainerDied","Data":"d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb"} Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.322721 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rlg4f" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.322764 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rlg4f" event={"ID":"2bfcec28-689c-4778-b9a6-2534ffb368b7","Type":"ContainerDied","Data":"3894c5b8ff1b903f962f93baa4ac4740f6620b3e4a9024be3d03da18e0ee5c97"} Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.322786 4715 scope.go:117] "RemoveContainer" containerID="d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.322797 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tpr6h" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerName="registry-server" containerID="cri-o://74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56" gracePeriod=2 Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.366939 4715 scope.go:117] "RemoveContainer" containerID="fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.379301 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rlg4f"] Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.387997 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rlg4f"] Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.403070 4715 scope.go:117] "RemoveContainer" containerID="db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.565223 4715 scope.go:117] "RemoveContainer" containerID="d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb" Dec 03 22:23:53 crc kubenswrapper[4715]: E1203 22:23:53.566944 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb\": container with ID starting with d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb not found: ID does not exist" containerID="d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.567375 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb"} err="failed to get container status \"d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb\": rpc error: code = NotFound desc = could not find container \"d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb\": container with ID starting with d5fc1e81e8cdfc308e8e6b325b8541ce8cec328a1f6714e2fd833caf0d120abb not found: ID does not exist" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.567476 4715 scope.go:117] "RemoveContainer" containerID="fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe" Dec 03 22:23:53 crc kubenswrapper[4715]: E1203 22:23:53.568141 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe\": container with ID starting with fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe not found: ID does not exist" containerID="fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.568203 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe"} err="failed to get container status \"fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe\": rpc error: code = NotFound desc = could not find container \"fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe\": container with ID starting with fd425b8a8a13b7a9a573b4b54002e769e9283ceacd08e55a80c43cd3f1bc2dbe not found: ID does not exist" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.568231 4715 scope.go:117] "RemoveContainer" containerID="db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f" Dec 03 22:23:53 crc kubenswrapper[4715]: E1203 22:23:53.568729 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f\": container with ID starting with db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f not found: ID does not exist" containerID="db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.568765 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f"} err="failed to get container status \"db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f\": rpc error: code = NotFound desc = could not find container \"db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f\": container with ID starting with db4447175772e62e6dec6da959412f5b64a3249fe93b329020ae3f2cac3e407f not found: ID does not exist" Dec 03 22:23:53 crc kubenswrapper[4715]: I1203 22:23:53.649405 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" path="/var/lib/kubelet/pods/2bfcec28-689c-4778-b9a6-2534ffb368b7/volumes" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.253535 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.334637 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerID="74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56" exitCode=0 Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.334707 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpr6h" event={"ID":"0ac36e60-4df8-4a07-8341-a77eb105e998","Type":"ContainerDied","Data":"74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56"} Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.334760 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tpr6h" event={"ID":"0ac36e60-4df8-4a07-8341-a77eb105e998","Type":"ContainerDied","Data":"3da4a11b787231d9756c12d69714020e601df3bfa49d1c2017254329d69a777b"} Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.334749 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tpr6h" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.334779 4715 scope.go:117] "RemoveContainer" containerID="74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.368674 4715 scope.go:117] "RemoveContainer" containerID="30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.377322 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqwlf\" (UniqueName: \"kubernetes.io/projected/0ac36e60-4df8-4a07-8341-a77eb105e998-kube-api-access-lqwlf\") pod \"0ac36e60-4df8-4a07-8341-a77eb105e998\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.377408 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-utilities\") pod \"0ac36e60-4df8-4a07-8341-a77eb105e998\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.377649 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-catalog-content\") pod \"0ac36e60-4df8-4a07-8341-a77eb105e998\" (UID: \"0ac36e60-4df8-4a07-8341-a77eb105e998\") " Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.379045 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-utilities" (OuterVolumeSpecName: "utilities") pod "0ac36e60-4df8-4a07-8341-a77eb105e998" (UID: "0ac36e60-4df8-4a07-8341-a77eb105e998"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.383168 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac36e60-4df8-4a07-8341-a77eb105e998-kube-api-access-lqwlf" (OuterVolumeSpecName: "kube-api-access-lqwlf") pod "0ac36e60-4df8-4a07-8341-a77eb105e998" (UID: "0ac36e60-4df8-4a07-8341-a77eb105e998"). InnerVolumeSpecName "kube-api-access-lqwlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.395106 4715 scope.go:117] "RemoveContainer" containerID="c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.408232 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ac36e60-4df8-4a07-8341-a77eb105e998" (UID: "0ac36e60-4df8-4a07-8341-a77eb105e998"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.428925 4715 scope.go:117] "RemoveContainer" containerID="74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56" Dec 03 22:23:54 crc kubenswrapper[4715]: E1203 22:23:54.429330 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56\": container with ID starting with 74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56 not found: ID does not exist" containerID="74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.429371 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56"} err="failed to get container status \"74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56\": rpc error: code = NotFound desc = could not find container \"74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56\": container with ID starting with 74821f801924958e825e6ed5e545a077eee0fddd1078ad2bc4e219e9fdd9fb56 not found: ID does not exist" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.429399 4715 scope.go:117] "RemoveContainer" containerID="30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53" Dec 03 22:23:54 crc kubenswrapper[4715]: E1203 22:23:54.429842 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53\": container with ID starting with 30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53 not found: ID does not exist" containerID="30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.429876 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53"} err="failed to get container status \"30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53\": rpc error: code = NotFound desc = could not find container \"30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53\": container with ID starting with 30b6e19dc95f0d20a1ccb51ac4cb1c056cb63e0d54cfd41501e9f55e1283ec53 not found: ID does not exist" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.429893 4715 scope.go:117] "RemoveContainer" containerID="c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d" Dec 03 22:23:54 crc kubenswrapper[4715]: E1203 22:23:54.430110 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d\": container with ID starting with c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d not found: ID does not exist" containerID="c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.430133 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d"} err="failed to get container status \"c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d\": rpc error: code = NotFound desc = could not find container \"c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d\": container with ID starting with c8cdadeab4053861d61d47b718a3c664cf405270d6ef1cbd2f0b8e9006e7ea4d not found: ID does not exist" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.480803 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.480856 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqwlf\" (UniqueName: \"kubernetes.io/projected/0ac36e60-4df8-4a07-8341-a77eb105e998-kube-api-access-lqwlf\") on node \"crc\" DevicePath \"\"" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.480879 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac36e60-4df8-4a07-8341-a77eb105e998-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.678340 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpr6h"] Dec 03 22:23:54 crc kubenswrapper[4715]: I1203 22:23:54.688711 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tpr6h"] Dec 03 22:23:55 crc kubenswrapper[4715]: I1203 22:23:55.646648 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" path="/var/lib/kubelet/pods/0ac36e60-4df8-4a07-8341-a77eb105e998/volumes" Dec 03 22:24:04 crc kubenswrapper[4715]: I1203 22:24:04.635048 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:24:04 crc kubenswrapper[4715]: E1203 22:24:04.635963 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:24:17 crc kubenswrapper[4715]: I1203 22:24:17.634433 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:24:20 crc kubenswrapper[4715]: I1203 22:24:20.600592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"33ce01d1db2f2fd17a93394b524ae7fd388810ef368375808b76327aec02d590"} Dec 03 22:25:06 crc kubenswrapper[4715]: I1203 22:25:06.070314 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-gc4h7"] Dec 03 22:25:06 crc kubenswrapper[4715]: I1203 22:25:06.082460 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-gc4h7"] Dec 03 22:25:06 crc kubenswrapper[4715]: I1203 22:25:06.092113 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2344-account-create-update-2zxn7"] Dec 03 22:25:06 crc kubenswrapper[4715]: I1203 22:25:06.102303 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2344-account-create-update-2zxn7"] Dec 03 22:25:07 crc kubenswrapper[4715]: I1203 22:25:07.029113 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-9js9r"] Dec 03 22:25:07 crc kubenswrapper[4715]: I1203 22:25:07.039954 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-9js9r"] Dec 03 22:25:07 crc kubenswrapper[4715]: I1203 22:25:07.052145 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-1c8c-account-create-update-8rqw4"] Dec 03 22:25:07 crc kubenswrapper[4715]: I1203 22:25:07.062259 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-1c8c-account-create-update-8rqw4"] Dec 03 22:25:07 crc kubenswrapper[4715]: I1203 22:25:07.647590 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a676b7b-101e-4fd5-8aae-828dc3c502a7" path="/var/lib/kubelet/pods/4a676b7b-101e-4fd5-8aae-828dc3c502a7/volumes" Dec 03 22:25:07 crc kubenswrapper[4715]: I1203 22:25:07.648966 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ab729b6-f4d1-4f48-bb28-43d4c4c81e58" path="/var/lib/kubelet/pods/4ab729b6-f4d1-4f48-bb28-43d4c4c81e58/volumes" Dec 03 22:25:07 crc kubenswrapper[4715]: I1203 22:25:07.650133 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdab221f-c97f-45d1-992a-d00b1e862b1a" path="/var/lib/kubelet/pods/bdab221f-c97f-45d1-992a-d00b1e862b1a/volumes" Dec 03 22:25:07 crc kubenswrapper[4715]: I1203 22:25:07.651305 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe6fd2a-8278-4fce-b29f-2f2237c5abf6" path="/var/lib/kubelet/pods/efe6fd2a-8278-4fce-b29f-2f2237c5abf6/volumes" Dec 03 22:25:17 crc kubenswrapper[4715]: I1203 22:25:17.045539 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-eeb1-account-create-update-hg4zn"] Dec 03 22:25:17 crc kubenswrapper[4715]: I1203 22:25:17.058350 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-9l4gg"] Dec 03 22:25:17 crc kubenswrapper[4715]: I1203 22:25:17.071849 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-eeb1-account-create-update-hg4zn"] Dec 03 22:25:17 crc kubenswrapper[4715]: I1203 22:25:17.084322 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-9l4gg"] Dec 03 22:25:17 crc kubenswrapper[4715]: I1203 22:25:17.662603 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5046d57b-4ce3-4f2c-8cbc-0dee5590b458" path="/var/lib/kubelet/pods/5046d57b-4ce3-4f2c-8cbc-0dee5590b458/volumes" Dec 03 22:25:17 crc kubenswrapper[4715]: I1203 22:25:17.663482 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5872f688-95bb-4443-ba2d-f55ca9b05e97" path="/var/lib/kubelet/pods/5872f688-95bb-4443-ba2d-f55ca9b05e97/volumes" Dec 03 22:25:37 crc kubenswrapper[4715]: I1203 22:25:37.421352 4715 generic.go:334] "Generic (PLEG): container finished" podID="7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63" containerID="e9dd62558b17ee3cffb973e0885cfe2a0fce5de816f694f2fa8b834ecd04b851" exitCode=0 Dec 03 22:25:37 crc kubenswrapper[4715]: I1203 22:25:37.421444 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" event={"ID":"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63","Type":"ContainerDied","Data":"e9dd62558b17ee3cffb973e0885cfe2a0fce5de816f694f2fa8b834ecd04b851"} Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.589343 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.706806 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlk5z\" (UniqueName: \"kubernetes.io/projected/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-kube-api-access-zlk5z\") pod \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.706896 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-bootstrap-combined-ca-bundle\") pod \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.706975 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-ssh-key\") pod \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.707076 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-inventory\") pod \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\" (UID: \"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63\") " Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.712888 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63" (UID: "7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.713034 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-kube-api-access-zlk5z" (OuterVolumeSpecName: "kube-api-access-zlk5z") pod "7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63" (UID: "7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63"). InnerVolumeSpecName "kube-api-access-zlk5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.739930 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63" (UID: "7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.743714 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-inventory" (OuterVolumeSpecName: "inventory") pod "7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63" (UID: "7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.809361 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlk5z\" (UniqueName: \"kubernetes.io/projected/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-kube-api-access-zlk5z\") on node \"crc\" DevicePath \"\"" Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.809396 4715 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.809409 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:25:39 crc kubenswrapper[4715]: I1203 22:25:39.809418 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.043923 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-8cfqn"] Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.063327 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-8cfqn"] Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.457577 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" event={"ID":"7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63","Type":"ContainerDied","Data":"de01b7c929faa1e76be901f6ff51a977f8779a66e4e806b38fe615bb100fcf0f"} Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.457624 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de01b7c929faa1e76be901f6ff51a977f8779a66e4e806b38fe615bb100fcf0f" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.457700 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.708688 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5"] Dec 03 22:25:40 crc kubenswrapper[4715]: E1203 22:25:40.709608 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerName="registry-server" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.709644 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerName="registry-server" Dec 03 22:25:40 crc kubenswrapper[4715]: E1203 22:25:40.709694 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerName="extract-utilities" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.709708 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerName="extract-utilities" Dec 03 22:25:40 crc kubenswrapper[4715]: E1203 22:25:40.709737 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerName="extract-content" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.709751 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerName="extract-content" Dec 03 22:25:40 crc kubenswrapper[4715]: E1203 22:25:40.709774 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerName="registry-server" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.709786 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerName="registry-server" Dec 03 22:25:40 crc kubenswrapper[4715]: E1203 22:25:40.709822 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerName="extract-content" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.709835 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerName="extract-content" Dec 03 22:25:40 crc kubenswrapper[4715]: E1203 22:25:40.709864 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerName="extract-utilities" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.709882 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerName="extract-utilities" Dec 03 22:25:40 crc kubenswrapper[4715]: E1203 22:25:40.709906 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.709923 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.710637 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac36e60-4df8-4a07-8341-a77eb105e998" containerName="registry-server" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.710679 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.710718 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bfcec28-689c-4778-b9a6-2534ffb368b7" containerName="registry-server" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.711864 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.714179 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.714901 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.716290 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.718080 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.720661 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5"] Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.837810 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.838018 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.838105 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4hdw\" (UniqueName: \"kubernetes.io/projected/54307606-6028-435b-bb67-3dd55f7a938a-kube-api-access-w4hdw\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.940138 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.940223 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4hdw\" (UniqueName: \"kubernetes.io/projected/54307606-6028-435b-bb67-3dd55f7a938a-kube-api-access-w4hdw\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.940261 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.947425 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.948129 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:40 crc kubenswrapper[4715]: I1203 22:25:40.971692 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4hdw\" (UniqueName: \"kubernetes.io/projected/54307606-6028-435b-bb67-3dd55f7a938a-kube-api-access-w4hdw\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:41 crc kubenswrapper[4715]: I1203 22:25:41.032867 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:25:41 crc kubenswrapper[4715]: I1203 22:25:41.573972 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5"] Dec 03 22:25:41 crc kubenswrapper[4715]: I1203 22:25:41.647006 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="423716d6-494f-429a-98b4-696e1387bbc3" path="/var/lib/kubelet/pods/423716d6-494f-429a-98b4-696e1387bbc3/volumes" Dec 03 22:25:42 crc kubenswrapper[4715]: I1203 22:25:42.482727 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" event={"ID":"54307606-6028-435b-bb67-3dd55f7a938a","Type":"ContainerStarted","Data":"6195dde23a7143343362806313489ca69317595c4e2afab01678d04b8c110b33"} Dec 03 22:25:44 crc kubenswrapper[4715]: I1203 22:25:44.435585 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:25:45 crc kubenswrapper[4715]: I1203 22:25:45.515897 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" event={"ID":"54307606-6028-435b-bb67-3dd55f7a938a","Type":"ContainerStarted","Data":"d4327e87e0f5c51fbec940205fdc32f0a892580363b029065fe090fa060da894"} Dec 03 22:25:45 crc kubenswrapper[4715]: I1203 22:25:45.542068 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" podStartSLOduration=2.68465259 podStartE2EDuration="5.542050295s" podCreationTimestamp="2025-12-03 22:25:40 +0000 UTC" firstStartedPulling="2025-12-03 22:25:41.575095658 +0000 UTC m=+2638.317806293" lastFinishedPulling="2025-12-03 22:25:44.432493363 +0000 UTC m=+2641.175203998" observedRunningTime="2025-12-03 22:25:45.531171774 +0000 UTC m=+2642.273882369" watchObservedRunningTime="2025-12-03 22:25:45.542050295 +0000 UTC m=+2642.284760890" Dec 03 22:25:53 crc kubenswrapper[4715]: I1203 22:25:53.610961 4715 scope.go:117] "RemoveContainer" containerID="b35a24b8a7c11936af70b902121801d8d56caadfb1006683f835e78a475872fe" Dec 03 22:25:53 crc kubenswrapper[4715]: I1203 22:25:53.648606 4715 scope.go:117] "RemoveContainer" containerID="810ee17edc7caf608374db8da1fc528a65b67b80c0a831e8b543e7649025f8e5" Dec 03 22:25:53 crc kubenswrapper[4715]: I1203 22:25:53.712539 4715 scope.go:117] "RemoveContainer" containerID="185c002ca496dccc6f7008715543c806b0ab76d68547ba111e9dbd24de2cf8cd" Dec 03 22:25:53 crc kubenswrapper[4715]: I1203 22:25:53.770131 4715 scope.go:117] "RemoveContainer" containerID="2c508d00b0d0d5fbc4e608b3df2d31feb8cc94c11cc4ce0a099ed3beea3aba97" Dec 03 22:25:53 crc kubenswrapper[4715]: I1203 22:25:53.804317 4715 scope.go:117] "RemoveContainer" containerID="77231f7728a88f0edb7e29432c877cd91f0f46986c5b237eef2a47a80c138f42" Dec 03 22:25:53 crc kubenswrapper[4715]: I1203 22:25:53.845307 4715 scope.go:117] "RemoveContainer" containerID="0b2a39c3654d7bea09c4da32a22208bf980028daf9b82fa49d48e4d8cf048b2d" Dec 03 22:25:53 crc kubenswrapper[4715]: I1203 22:25:53.877868 4715 scope.go:117] "RemoveContainer" containerID="4c59cd8db222e8055fd4a086fb0bf9cd2d0bde2960ada2dddcd98d8b19da62d7" Dec 03 22:25:53 crc kubenswrapper[4715]: I1203 22:25:53.921613 4715 scope.go:117] "RemoveContainer" containerID="98b64c6fa621fe81b53455bf9c664b08983758898a8640904a945e9e338a7868" Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.066963 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-j4dtp"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.087395 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-hg7kd"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.096314 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-tzppt"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.106123 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-fd69-account-create-update-n7kqs"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.120056 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-de87-account-create-update-4tn28"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.131610 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-fd69-account-create-update-n7kqs"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.139964 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-j4dtp"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.149439 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-tzppt"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.158465 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-de87-account-create-update-4tn28"] Dec 03 22:25:54 crc kubenswrapper[4715]: I1203 22:25:54.170276 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-hg7kd"] Dec 03 22:25:55 crc kubenswrapper[4715]: I1203 22:25:55.667905 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="075e50c6-4819-42c7-88b3-a73a7f959f81" path="/var/lib/kubelet/pods/075e50c6-4819-42c7-88b3-a73a7f959f81/volumes" Dec 03 22:25:55 crc kubenswrapper[4715]: I1203 22:25:55.669526 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60b5688d-8ef1-44f9-b05e-60e99834e994" path="/var/lib/kubelet/pods/60b5688d-8ef1-44f9-b05e-60e99834e994/volumes" Dec 03 22:25:55 crc kubenswrapper[4715]: I1203 22:25:55.670749 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75b6412c-5008-4966-b1d9-55f9846d9cd1" path="/var/lib/kubelet/pods/75b6412c-5008-4966-b1d9-55f9846d9cd1/volumes" Dec 03 22:25:55 crc kubenswrapper[4715]: I1203 22:25:55.672614 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4" path="/var/lib/kubelet/pods/78c3f6f3-ad41-4173-ad77-b6b2fa76d6b4/volumes" Dec 03 22:25:55 crc kubenswrapper[4715]: I1203 22:25:55.673818 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7" path="/var/lib/kubelet/pods/f2e31ed8-58e8-4ee3-a7d7-ee6f8dec6ab7/volumes" Dec 03 22:26:03 crc kubenswrapper[4715]: I1203 22:26:03.045396 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-e283-account-create-update-r7wc9"] Dec 03 22:26:03 crc kubenswrapper[4715]: I1203 22:26:03.055852 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-e283-account-create-update-r7wc9"] Dec 03 22:26:03 crc kubenswrapper[4715]: I1203 22:26:03.650994 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79758367-7636-4dcd-9026-b4008520c52d" path="/var/lib/kubelet/pods/79758367-7636-4dcd-9026-b4008520c52d/volumes" Dec 03 22:26:08 crc kubenswrapper[4715]: I1203 22:26:08.032599 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-d7jfj"] Dec 03 22:26:08 crc kubenswrapper[4715]: I1203 22:26:08.042975 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-d7jfj"] Dec 03 22:26:09 crc kubenswrapper[4715]: I1203 22:26:09.649151 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e7668eb-91d3-4172-9b37-048835ae6899" path="/var/lib/kubelet/pods/7e7668eb-91d3-4172-9b37-048835ae6899/volumes" Dec 03 22:26:35 crc kubenswrapper[4715]: I1203 22:26:35.159842 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:26:35 crc kubenswrapper[4715]: I1203 22:26:35.160417 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:26:52 crc kubenswrapper[4715]: I1203 22:26:52.065288 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-blfqt"] Dec 03 22:26:52 crc kubenswrapper[4715]: I1203 22:26:52.084269 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-blfqt"] Dec 03 22:26:53 crc kubenswrapper[4715]: I1203 22:26:53.044987 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-d8zcz"] Dec 03 22:26:53 crc kubenswrapper[4715]: I1203 22:26:53.059665 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-d8zcz"] Dec 03 22:26:53 crc kubenswrapper[4715]: I1203 22:26:53.655975 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ded23d0-fc8b-4ed7-8358-573da1f24919" path="/var/lib/kubelet/pods/6ded23d0-fc8b-4ed7-8358-573da1f24919/volumes" Dec 03 22:26:53 crc kubenswrapper[4715]: I1203 22:26:53.657556 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3e2fa00-220e-4811-8b41-644d96b70a46" path="/var/lib/kubelet/pods/d3e2fa00-220e-4811-8b41-644d96b70a46/volumes" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.130543 4715 scope.go:117] "RemoveContainer" containerID="371aa80cbef5669afbd1d3dc6067b637a7c1618fca9e34f0e35de061c608b7c8" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.194386 4715 scope.go:117] "RemoveContainer" containerID="3fa789e13ceb6805ffdd0b41900fc86eed3857e22eaf5896774011c906deb354" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.242639 4715 scope.go:117] "RemoveContainer" containerID="535f40006dbf7561ccd0f094557951784b4e3ae0e26192d9b1ab42c3bcc862a2" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.294252 4715 scope.go:117] "RemoveContainer" containerID="55cf90793dbe13f14c95414aed7d758928f497a82c05bab05454946788d83cf9" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.320551 4715 scope.go:117] "RemoveContainer" containerID="0b33a3889d97525c66bbdfebd8e359f643585e31acd484a171e876472aeb31f1" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.361860 4715 scope.go:117] "RemoveContainer" containerID="346469aa378873aea0de480d2af9b2f14c2fd22f9ca74f327dcae8f8fadab8cf" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.404557 4715 scope.go:117] "RemoveContainer" containerID="68962d8f1e4c9519115b629e15aa9a5c90993e36f0d3dc9b74e72ad203cbb42d" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.425708 4715 scope.go:117] "RemoveContainer" containerID="d2bd265d54c25da8c6955025d08aa8864ef8690abc4a1968a9b1864bbd709c7d" Dec 03 22:26:54 crc kubenswrapper[4715]: I1203 22:26:54.448956 4715 scope.go:117] "RemoveContainer" containerID="54ea093b6c7ea48d7d2d786ef880b3512293ff422760c010dd224c8392081bae" Dec 03 22:27:01 crc kubenswrapper[4715]: I1203 22:27:01.044570 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-kp478"] Dec 03 22:27:01 crc kubenswrapper[4715]: I1203 22:27:01.055519 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-kp478"] Dec 03 22:27:01 crc kubenswrapper[4715]: I1203 22:27:01.649717 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60edd1a1-9b8a-4d7d-977b-11c834922ca1" path="/var/lib/kubelet/pods/60edd1a1-9b8a-4d7d-977b-11c834922ca1/volumes" Dec 03 22:27:02 crc kubenswrapper[4715]: I1203 22:27:02.039030 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-6ngmv"] Dec 03 22:27:02 crc kubenswrapper[4715]: I1203 22:27:02.054944 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-6ngmv"] Dec 03 22:27:03 crc kubenswrapper[4715]: I1203 22:27:03.671832 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2eba90e4-4991-4260-9e8e-2a3a5412b264" path="/var/lib/kubelet/pods/2eba90e4-4991-4260-9e8e-2a3a5412b264/volumes" Dec 03 22:27:05 crc kubenswrapper[4715]: I1203 22:27:05.159812 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:27:05 crc kubenswrapper[4715]: I1203 22:27:05.160130 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:27:21 crc kubenswrapper[4715]: I1203 22:27:21.050563 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-p4mjk"] Dec 03 22:27:21 crc kubenswrapper[4715]: I1203 22:27:21.068995 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-p4mjk"] Dec 03 22:27:21 crc kubenswrapper[4715]: I1203 22:27:21.647216 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ec32c20-9117-4eca-9b73-a7f8b84677da" path="/var/lib/kubelet/pods/5ec32c20-9117-4eca-9b73-a7f8b84677da/volumes" Dec 03 22:27:35 crc kubenswrapper[4715]: I1203 22:27:35.159696 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:27:35 crc kubenswrapper[4715]: I1203 22:27:35.160381 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:27:35 crc kubenswrapper[4715]: I1203 22:27:35.160423 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:27:35 crc kubenswrapper[4715]: I1203 22:27:35.161123 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33ce01d1db2f2fd17a93394b524ae7fd388810ef368375808b76327aec02d590"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:27:35 crc kubenswrapper[4715]: I1203 22:27:35.161179 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://33ce01d1db2f2fd17a93394b524ae7fd388810ef368375808b76327aec02d590" gracePeriod=600 Dec 03 22:27:36 crc kubenswrapper[4715]: I1203 22:27:36.047073 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="33ce01d1db2f2fd17a93394b524ae7fd388810ef368375808b76327aec02d590" exitCode=0 Dec 03 22:27:36 crc kubenswrapper[4715]: I1203 22:27:36.047165 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"33ce01d1db2f2fd17a93394b524ae7fd388810ef368375808b76327aec02d590"} Dec 03 22:27:36 crc kubenswrapper[4715]: I1203 22:27:36.047438 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a"} Dec 03 22:27:36 crc kubenswrapper[4715]: I1203 22:27:36.047463 4715 scope.go:117] "RemoveContainer" containerID="ebef3be4b6aba481ba20c5f76091454aee4c283b40584dc81cfc85eb23bb21cc" Dec 03 22:27:38 crc kubenswrapper[4715]: I1203 22:27:38.064772 4715 generic.go:334] "Generic (PLEG): container finished" podID="54307606-6028-435b-bb67-3dd55f7a938a" containerID="d4327e87e0f5c51fbec940205fdc32f0a892580363b029065fe090fa060da894" exitCode=0 Dec 03 22:27:38 crc kubenswrapper[4715]: I1203 22:27:38.064848 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" event={"ID":"54307606-6028-435b-bb67-3dd55f7a938a","Type":"ContainerDied","Data":"d4327e87e0f5c51fbec940205fdc32f0a892580363b029065fe090fa060da894"} Dec 03 22:27:39 crc kubenswrapper[4715]: I1203 22:27:39.646528 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:27:39 crc kubenswrapper[4715]: I1203 22:27:39.791547 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-inventory\") pod \"54307606-6028-435b-bb67-3dd55f7a938a\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " Dec 03 22:27:39 crc kubenswrapper[4715]: I1203 22:27:39.792172 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4hdw\" (UniqueName: \"kubernetes.io/projected/54307606-6028-435b-bb67-3dd55f7a938a-kube-api-access-w4hdw\") pod \"54307606-6028-435b-bb67-3dd55f7a938a\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " Dec 03 22:27:39 crc kubenswrapper[4715]: I1203 22:27:39.792302 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-ssh-key\") pod \"54307606-6028-435b-bb67-3dd55f7a938a\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.089745 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54307606-6028-435b-bb67-3dd55f7a938a-kube-api-access-w4hdw" (OuterVolumeSpecName: "kube-api-access-w4hdw") pod "54307606-6028-435b-bb67-3dd55f7a938a" (UID: "54307606-6028-435b-bb67-3dd55f7a938a"). InnerVolumeSpecName "kube-api-access-w4hdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.091008 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" event={"ID":"54307606-6028-435b-bb67-3dd55f7a938a","Type":"ContainerDied","Data":"6195dde23a7143343362806313489ca69317595c4e2afab01678d04b8c110b33"} Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.091075 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6195dde23a7143343362806313489ca69317595c4e2afab01678d04b8c110b33" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.091148 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.099748 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "54307606-6028-435b-bb67-3dd55f7a938a" (UID: "54307606-6028-435b-bb67-3dd55f7a938a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.099796 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-inventory" (OuterVolumeSpecName: "inventory") pod "54307606-6028-435b-bb67-3dd55f7a938a" (UID: "54307606-6028-435b-bb67-3dd55f7a938a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:27:40 crc kubenswrapper[4715]: W1203 22:27:40.100674 4715 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/54307606-6028-435b-bb67-3dd55f7a938a/volumes/kubernetes.io~secret/ssh-key Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.100698 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "54307606-6028-435b-bb67-3dd55f7a938a" (UID: "54307606-6028-435b-bb67-3dd55f7a938a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.100676 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-ssh-key\") pod \"54307606-6028-435b-bb67-3dd55f7a938a\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.100762 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-inventory\") pod \"54307606-6028-435b-bb67-3dd55f7a938a\" (UID: \"54307606-6028-435b-bb67-3dd55f7a938a\") " Dec 03 22:27:40 crc kubenswrapper[4715]: W1203 22:27:40.100862 4715 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/54307606-6028-435b-bb67-3dd55f7a938a/volumes/kubernetes.io~secret/inventory Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.100873 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-inventory" (OuterVolumeSpecName: "inventory") pod "54307606-6028-435b-bb67-3dd55f7a938a" (UID: "54307606-6028-435b-bb67-3dd55f7a938a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.101229 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4hdw\" (UniqueName: \"kubernetes.io/projected/54307606-6028-435b-bb67-3dd55f7a938a-kube-api-access-w4hdw\") on node \"crc\" DevicePath \"\"" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.101256 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.101268 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/54307606-6028-435b-bb67-3dd55f7a938a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.182197 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg"] Dec 03 22:27:40 crc kubenswrapper[4715]: E1203 22:27:40.182738 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54307606-6028-435b-bb67-3dd55f7a938a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.182762 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="54307606-6028-435b-bb67-3dd55f7a938a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.182990 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="54307606-6028-435b-bb67-3dd55f7a938a" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.183760 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.192593 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg"] Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.309014 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.309104 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvslg\" (UniqueName: \"kubernetes.io/projected/7fa5989a-f985-4757-84dd-3c658472a85e-kube-api-access-mvslg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.309437 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.410810 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.410884 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.410917 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvslg\" (UniqueName: \"kubernetes.io/projected/7fa5989a-f985-4757-84dd-3c658472a85e-kube-api-access-mvslg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.414057 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.414458 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.427876 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvslg\" (UniqueName: \"kubernetes.io/projected/7fa5989a-f985-4757-84dd-3c658472a85e-kube-api-access-mvslg\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:40 crc kubenswrapper[4715]: I1203 22:27:40.517751 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:27:41 crc kubenswrapper[4715]: I1203 22:27:41.107616 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg"] Dec 03 22:27:41 crc kubenswrapper[4715]: I1203 22:27:41.113329 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:27:42 crc kubenswrapper[4715]: I1203 22:27:42.113562 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" event={"ID":"7fa5989a-f985-4757-84dd-3c658472a85e","Type":"ContainerStarted","Data":"47a5bfcb4564e7b016f594b267b4af2a3bcae1093c6903d95258fdd90c636436"} Dec 03 22:27:42 crc kubenswrapper[4715]: I1203 22:27:42.113889 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" event={"ID":"7fa5989a-f985-4757-84dd-3c658472a85e","Type":"ContainerStarted","Data":"d88336b504d068424eff62e486280dad23908a1b13c031d44a5527dcaf61975e"} Dec 03 22:27:42 crc kubenswrapper[4715]: I1203 22:27:42.141259 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" podStartSLOduration=1.7337352419999998 podStartE2EDuration="2.141214066s" podCreationTimestamp="2025-12-03 22:27:40 +0000 UTC" firstStartedPulling="2025-12-03 22:27:41.113065554 +0000 UTC m=+2757.855776159" lastFinishedPulling="2025-12-03 22:27:41.520544388 +0000 UTC m=+2758.263254983" observedRunningTime="2025-12-03 22:27:42.132628566 +0000 UTC m=+2758.875339151" watchObservedRunningTime="2025-12-03 22:27:42.141214066 +0000 UTC m=+2758.883924701" Dec 03 22:27:54 crc kubenswrapper[4715]: I1203 22:27:54.618588 4715 scope.go:117] "RemoveContainer" containerID="ad4f3a024bad21cc3c9bcda15545d8af611a2f98e124ad9fcb76eb260681987f" Dec 03 22:27:54 crc kubenswrapper[4715]: I1203 22:27:54.669231 4715 scope.go:117] "RemoveContainer" containerID="16fd66c726542fba27a9cf3c4da98d61f54d7099995437d0013b10cd48184d2a" Dec 03 22:27:54 crc kubenswrapper[4715]: I1203 22:27:54.717864 4715 scope.go:117] "RemoveContainer" containerID="8b4284641a08341e55d09f63043ef7d4f7d7d31499b72a5302b241474ad97d4a" Dec 03 22:28:00 crc kubenswrapper[4715]: I1203 22:28:00.775389 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9lgs7"] Dec 03 22:28:00 crc kubenswrapper[4715]: I1203 22:28:00.778157 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:00 crc kubenswrapper[4715]: I1203 22:28:00.795410 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lgs7"] Dec 03 22:28:00 crc kubenswrapper[4715]: I1203 22:28:00.901756 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-catalog-content\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:00 crc kubenswrapper[4715]: I1203 22:28:00.902092 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-utilities\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:00 crc kubenswrapper[4715]: I1203 22:28:00.902150 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf4pt\" (UniqueName: \"kubernetes.io/projected/fd2bb600-649f-4a1f-8126-aca8fd4f785f-kube-api-access-hf4pt\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:01 crc kubenswrapper[4715]: I1203 22:28:01.003982 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf4pt\" (UniqueName: \"kubernetes.io/projected/fd2bb600-649f-4a1f-8126-aca8fd4f785f-kube-api-access-hf4pt\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:01 crc kubenswrapper[4715]: I1203 22:28:01.004174 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-catalog-content\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:01 crc kubenswrapper[4715]: I1203 22:28:01.004244 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-utilities\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:01 crc kubenswrapper[4715]: I1203 22:28:01.004943 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-catalog-content\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:01 crc kubenswrapper[4715]: I1203 22:28:01.004986 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-utilities\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:01 crc kubenswrapper[4715]: I1203 22:28:01.025390 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf4pt\" (UniqueName: \"kubernetes.io/projected/fd2bb600-649f-4a1f-8126-aca8fd4f785f-kube-api-access-hf4pt\") pod \"certified-operators-9lgs7\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:01 crc kubenswrapper[4715]: I1203 22:28:01.098621 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:01 crc kubenswrapper[4715]: I1203 22:28:01.627793 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lgs7"] Dec 03 22:28:01 crc kubenswrapper[4715]: W1203 22:28:01.628493 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd2bb600_649f_4a1f_8126_aca8fd4f785f.slice/crio-30350d72f06cdf36f1012d250511943d50a3ddd2b3d11dbb85a18038b4112ca5 WatchSource:0}: Error finding container 30350d72f06cdf36f1012d250511943d50a3ddd2b3d11dbb85a18038b4112ca5: Status 404 returned error can't find the container with id 30350d72f06cdf36f1012d250511943d50a3ddd2b3d11dbb85a18038b4112ca5 Dec 03 22:28:02 crc kubenswrapper[4715]: I1203 22:28:02.322517 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lgs7" event={"ID":"fd2bb600-649f-4a1f-8126-aca8fd4f785f","Type":"ContainerStarted","Data":"30350d72f06cdf36f1012d250511943d50a3ddd2b3d11dbb85a18038b4112ca5"} Dec 03 22:28:03 crc kubenswrapper[4715]: I1203 22:28:03.331330 4715 generic.go:334] "Generic (PLEG): container finished" podID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerID="2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9" exitCode=0 Dec 03 22:28:03 crc kubenswrapper[4715]: I1203 22:28:03.331374 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lgs7" event={"ID":"fd2bb600-649f-4a1f-8126-aca8fd4f785f","Type":"ContainerDied","Data":"2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9"} Dec 03 22:28:04 crc kubenswrapper[4715]: I1203 22:28:04.343344 4715 generic.go:334] "Generic (PLEG): container finished" podID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerID="22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7" exitCode=0 Dec 03 22:28:04 crc kubenswrapper[4715]: I1203 22:28:04.343554 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lgs7" event={"ID":"fd2bb600-649f-4a1f-8126-aca8fd4f785f","Type":"ContainerDied","Data":"22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7"} Dec 03 22:28:05 crc kubenswrapper[4715]: I1203 22:28:05.359542 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lgs7" event={"ID":"fd2bb600-649f-4a1f-8126-aca8fd4f785f","Type":"ContainerStarted","Data":"a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7"} Dec 03 22:28:05 crc kubenswrapper[4715]: I1203 22:28:05.382019 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9lgs7" podStartSLOduration=3.948958164 podStartE2EDuration="5.381994841s" podCreationTimestamp="2025-12-03 22:28:00 +0000 UTC" firstStartedPulling="2025-12-03 22:28:03.333941927 +0000 UTC m=+2780.076652522" lastFinishedPulling="2025-12-03 22:28:04.766978604 +0000 UTC m=+2781.509689199" observedRunningTime="2025-12-03 22:28:05.379414142 +0000 UTC m=+2782.122124767" watchObservedRunningTime="2025-12-03 22:28:05.381994841 +0000 UTC m=+2782.124705456" Dec 03 22:28:11 crc kubenswrapper[4715]: I1203 22:28:11.099644 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:11 crc kubenswrapper[4715]: I1203 22:28:11.100171 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:11 crc kubenswrapper[4715]: I1203 22:28:11.155183 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:11 crc kubenswrapper[4715]: I1203 22:28:11.462485 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:11 crc kubenswrapper[4715]: I1203 22:28:11.510032 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lgs7"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.066858 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-4nhvz"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.078354 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-d43a-account-create-update-dscvc"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.089879 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-6ql9s"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.098628 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-181b-account-create-update-c86tf"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.105166 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-9nwdg"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.112068 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-6ql9s"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.119083 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-4nhvz"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.126798 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-46c1-account-create-update-vfm2h"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.133793 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-d43a-account-create-update-dscvc"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.140530 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-181b-account-create-update-c86tf"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.147250 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-9nwdg"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.153667 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-46c1-account-create-update-vfm2h"] Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.453588 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9lgs7" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerName="registry-server" containerID="cri-o://a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7" gracePeriod=2 Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.646405 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="094b2293-50c2-4755-8abe-798dc10adb24" path="/var/lib/kubelet/pods/094b2293-50c2-4755-8abe-798dc10adb24/volumes" Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.647247 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946473d7-6235-4989-b8b5-8bf32fac8873" path="/var/lib/kubelet/pods/946473d7-6235-4989-b8b5-8bf32fac8873/volumes" Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.647956 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a3f66d8-7bbb-4939-bef3-ebb7365e53a2" path="/var/lib/kubelet/pods/9a3f66d8-7bbb-4939-bef3-ebb7365e53a2/volumes" Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.648585 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba52bb7b-0acf-4954-b20c-34b35f5b0169" path="/var/lib/kubelet/pods/ba52bb7b-0acf-4954-b20c-34b35f5b0169/volumes" Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.649155 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c39bbab4-12df-47cb-b3f2-85e2d2f8cac9" path="/var/lib/kubelet/pods/c39bbab4-12df-47cb-b3f2-85e2d2f8cac9/volumes" Dec 03 22:28:13 crc kubenswrapper[4715]: I1203 22:28:13.649755 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9bd7ccb-3d1b-4621-89e8-96751e61be26" path="/var/lib/kubelet/pods/e9bd7ccb-3d1b-4621-89e8-96751e61be26/volumes" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.078708 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.262428 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf4pt\" (UniqueName: \"kubernetes.io/projected/fd2bb600-649f-4a1f-8126-aca8fd4f785f-kube-api-access-hf4pt\") pod \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.262795 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-catalog-content\") pod \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.262855 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-utilities\") pod \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\" (UID: \"fd2bb600-649f-4a1f-8126-aca8fd4f785f\") " Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.263547 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-utilities" (OuterVolumeSpecName: "utilities") pod "fd2bb600-649f-4a1f-8126-aca8fd4f785f" (UID: "fd2bb600-649f-4a1f-8126-aca8fd4f785f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.270618 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd2bb600-649f-4a1f-8126-aca8fd4f785f-kube-api-access-hf4pt" (OuterVolumeSpecName: "kube-api-access-hf4pt") pod "fd2bb600-649f-4a1f-8126-aca8fd4f785f" (UID: "fd2bb600-649f-4a1f-8126-aca8fd4f785f"). InnerVolumeSpecName "kube-api-access-hf4pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.326890 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd2bb600-649f-4a1f-8126-aca8fd4f785f" (UID: "fd2bb600-649f-4a1f-8126-aca8fd4f785f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.365677 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.365747 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd2bb600-649f-4a1f-8126-aca8fd4f785f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.365776 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf4pt\" (UniqueName: \"kubernetes.io/projected/fd2bb600-649f-4a1f-8126-aca8fd4f785f-kube-api-access-hf4pt\") on node \"crc\" DevicePath \"\"" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.463164 4715 generic.go:334] "Generic (PLEG): container finished" podID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerID="a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7" exitCode=0 Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.463206 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lgs7" event={"ID":"fd2bb600-649f-4a1f-8126-aca8fd4f785f","Type":"ContainerDied","Data":"a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7"} Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.463231 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lgs7" event={"ID":"fd2bb600-649f-4a1f-8126-aca8fd4f785f","Type":"ContainerDied","Data":"30350d72f06cdf36f1012d250511943d50a3ddd2b3d11dbb85a18038b4112ca5"} Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.463248 4715 scope.go:117] "RemoveContainer" containerID="a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.463356 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lgs7" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.500619 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lgs7"] Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.503654 4715 scope.go:117] "RemoveContainer" containerID="22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.516219 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9lgs7"] Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.523047 4715 scope.go:117] "RemoveContainer" containerID="2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.569436 4715 scope.go:117] "RemoveContainer" containerID="a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7" Dec 03 22:28:14 crc kubenswrapper[4715]: E1203 22:28:14.570073 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7\": container with ID starting with a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7 not found: ID does not exist" containerID="a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.570119 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7"} err="failed to get container status \"a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7\": rpc error: code = NotFound desc = could not find container \"a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7\": container with ID starting with a91d35fc12d35ca74376a392d3bf101192fbafe594a78c5f4f409e08fb5b1fe7 not found: ID does not exist" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.570169 4715 scope.go:117] "RemoveContainer" containerID="22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7" Dec 03 22:28:14 crc kubenswrapper[4715]: E1203 22:28:14.570575 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7\": container with ID starting with 22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7 not found: ID does not exist" containerID="22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.570606 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7"} err="failed to get container status \"22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7\": rpc error: code = NotFound desc = could not find container \"22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7\": container with ID starting with 22a214b1c5d6eae4f9c628958f7f8055010e740015970490ba3916fb880a1df7 not found: ID does not exist" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.570626 4715 scope.go:117] "RemoveContainer" containerID="2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9" Dec 03 22:28:14 crc kubenswrapper[4715]: E1203 22:28:14.570937 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9\": container with ID starting with 2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9 not found: ID does not exist" containerID="2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9" Dec 03 22:28:14 crc kubenswrapper[4715]: I1203 22:28:14.570981 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9"} err="failed to get container status \"2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9\": rpc error: code = NotFound desc = could not find container \"2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9\": container with ID starting with 2c4e8fcd27c119c1d2ae1b19a881d4b39247abe52261015cf7f30d31080d96b9 not found: ID does not exist" Dec 03 22:28:15 crc kubenswrapper[4715]: I1203 22:28:15.645986 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" path="/var/lib/kubelet/pods/fd2bb600-649f-4a1f-8126-aca8fd4f785f/volumes" Dec 03 22:28:44 crc kubenswrapper[4715]: I1203 22:28:44.049077 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m68zt"] Dec 03 22:28:44 crc kubenswrapper[4715]: I1203 22:28:44.061782 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-m68zt"] Dec 03 22:28:45 crc kubenswrapper[4715]: I1203 22:28:45.649195 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="008014fe-5f97-4de4-a7ef-4bf119aed38f" path="/var/lib/kubelet/pods/008014fe-5f97-4de4-a7ef-4bf119aed38f/volumes" Dec 03 22:28:54 crc kubenswrapper[4715]: I1203 22:28:54.833271 4715 scope.go:117] "RemoveContainer" containerID="42708c24c3bcec216fe4dd11b4853a7f1bd27b13321c4b02aa48f63a9130ef0d" Dec 03 22:28:54 crc kubenswrapper[4715]: I1203 22:28:54.882531 4715 scope.go:117] "RemoveContainer" containerID="9b14ea35554400954df1f9303ee8433d233d9452cb383b06cf583b15be7d7d8d" Dec 03 22:28:54 crc kubenswrapper[4715]: I1203 22:28:54.921395 4715 scope.go:117] "RemoveContainer" containerID="692ba571f2899b4dfe7b0e82b9945b9d217970de042398d4c450448f61b5c0dd" Dec 03 22:28:54 crc kubenswrapper[4715]: I1203 22:28:54.983267 4715 scope.go:117] "RemoveContainer" containerID="a4406c2da4437418f16f953d3850ca7a0bc3619e2d1189b88ed76dbe28c67e27" Dec 03 22:28:55 crc kubenswrapper[4715]: I1203 22:28:55.002356 4715 scope.go:117] "RemoveContainer" containerID="fb011ef848a00c81e2da1976439fb931a33b9cae1948a954793a73aa15301343" Dec 03 22:28:55 crc kubenswrapper[4715]: I1203 22:28:55.047318 4715 scope.go:117] "RemoveContainer" containerID="f606fae52baae301c32649dab26448703d133e0c6c55cf5480176ebea303d755" Dec 03 22:28:55 crc kubenswrapper[4715]: I1203 22:28:55.089683 4715 scope.go:117] "RemoveContainer" containerID="84743759692db45134fdbe7f3474a79c4ce13731927ce11d290a5e950822455e" Dec 03 22:28:59 crc kubenswrapper[4715]: I1203 22:28:59.940022 4715 generic.go:334] "Generic (PLEG): container finished" podID="7fa5989a-f985-4757-84dd-3c658472a85e" containerID="47a5bfcb4564e7b016f594b267b4af2a3bcae1093c6903d95258fdd90c636436" exitCode=0 Dec 03 22:28:59 crc kubenswrapper[4715]: I1203 22:28:59.940149 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" event={"ID":"7fa5989a-f985-4757-84dd-3c658472a85e","Type":"ContainerDied","Data":"47a5bfcb4564e7b016f594b267b4af2a3bcae1093c6903d95258fdd90c636436"} Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.446547 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.544943 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvslg\" (UniqueName: \"kubernetes.io/projected/7fa5989a-f985-4757-84dd-3c658472a85e-kube-api-access-mvslg\") pod \"7fa5989a-f985-4757-84dd-3c658472a85e\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.545049 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-inventory\") pod \"7fa5989a-f985-4757-84dd-3c658472a85e\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.545088 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-ssh-key\") pod \"7fa5989a-f985-4757-84dd-3c658472a85e\" (UID: \"7fa5989a-f985-4757-84dd-3c658472a85e\") " Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.550382 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fa5989a-f985-4757-84dd-3c658472a85e-kube-api-access-mvslg" (OuterVolumeSpecName: "kube-api-access-mvslg") pod "7fa5989a-f985-4757-84dd-3c658472a85e" (UID: "7fa5989a-f985-4757-84dd-3c658472a85e"). InnerVolumeSpecName "kube-api-access-mvslg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.575586 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7fa5989a-f985-4757-84dd-3c658472a85e" (UID: "7fa5989a-f985-4757-84dd-3c658472a85e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.581464 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-inventory" (OuterVolumeSpecName: "inventory") pod "7fa5989a-f985-4757-84dd-3c658472a85e" (UID: "7fa5989a-f985-4757-84dd-3c658472a85e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.647606 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvslg\" (UniqueName: \"kubernetes.io/projected/7fa5989a-f985-4757-84dd-3c658472a85e-kube-api-access-mvslg\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.647669 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.647679 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fa5989a-f985-4757-84dd-3c658472a85e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.959167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" event={"ID":"7fa5989a-f985-4757-84dd-3c658472a85e","Type":"ContainerDied","Data":"d88336b504d068424eff62e486280dad23908a1b13c031d44a5527dcaf61975e"} Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.959208 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d88336b504d068424eff62e486280dad23908a1b13c031d44a5527dcaf61975e" Dec 03 22:29:01 crc kubenswrapper[4715]: I1203 22:29:01.959270 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.059046 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2"] Dec 03 22:29:02 crc kubenswrapper[4715]: E1203 22:29:02.059702 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerName="registry-server" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.059786 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerName="registry-server" Dec 03 22:29:02 crc kubenswrapper[4715]: E1203 22:29:02.059858 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerName="extract-content" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.059919 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerName="extract-content" Dec 03 22:29:02 crc kubenswrapper[4715]: E1203 22:29:02.060027 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerName="extract-utilities" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.060111 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerName="extract-utilities" Dec 03 22:29:02 crc kubenswrapper[4715]: E1203 22:29:02.060201 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fa5989a-f985-4757-84dd-3c658472a85e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.060253 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fa5989a-f985-4757-84dd-3c658472a85e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.060470 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd2bb600-649f-4a1f-8126-aca8fd4f785f" containerName="registry-server" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.060572 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fa5989a-f985-4757-84dd-3c658472a85e" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.061255 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.064758 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.064927 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.065078 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.071914 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.074856 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2"] Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.157699 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.157744 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.157787 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t495\" (UniqueName: \"kubernetes.io/projected/db7da37a-b417-433c-86e9-a18446b4af27-kube-api-access-4t495\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.260008 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.260095 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.260188 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t495\" (UniqueName: \"kubernetes.io/projected/db7da37a-b417-433c-86e9-a18446b4af27-kube-api-access-4t495\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.265405 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.273031 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.281673 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t495\" (UniqueName: \"kubernetes.io/projected/db7da37a-b417-433c-86e9-a18446b4af27-kube-api-access-4t495\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.379197 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.724551 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2"] Dec 03 22:29:02 crc kubenswrapper[4715]: W1203 22:29:02.727544 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb7da37a_b417_433c_86e9_a18446b4af27.slice/crio-be22066caf6f802d7e17ab0b7951c27e47a267434b8366299274f1b976e97a4a WatchSource:0}: Error finding container be22066caf6f802d7e17ab0b7951c27e47a267434b8366299274f1b976e97a4a: Status 404 returned error can't find the container with id be22066caf6f802d7e17ab0b7951c27e47a267434b8366299274f1b976e97a4a Dec 03 22:29:02 crc kubenswrapper[4715]: I1203 22:29:02.966711 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" event={"ID":"db7da37a-b417-433c-86e9-a18446b4af27","Type":"ContainerStarted","Data":"be22066caf6f802d7e17ab0b7951c27e47a267434b8366299274f1b976e97a4a"} Dec 03 22:29:03 crc kubenswrapper[4715]: I1203 22:29:03.987804 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" event={"ID":"db7da37a-b417-433c-86e9-a18446b4af27","Type":"ContainerStarted","Data":"2708ba6a22837a148100380497dc32ec7e01e3f7eb3d1d39e0ab627386f849f5"} Dec 03 22:29:04 crc kubenswrapper[4715]: I1203 22:29:04.016941 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" podStartSLOduration=1.55356435 podStartE2EDuration="2.016918379s" podCreationTimestamp="2025-12-03 22:29:02 +0000 UTC" firstStartedPulling="2025-12-03 22:29:02.729644832 +0000 UTC m=+2839.472355427" lastFinishedPulling="2025-12-03 22:29:03.192998851 +0000 UTC m=+2839.935709456" observedRunningTime="2025-12-03 22:29:04.009202572 +0000 UTC m=+2840.751913187" watchObservedRunningTime="2025-12-03 22:29:04.016918379 +0000 UTC m=+2840.759628984" Dec 03 22:29:06 crc kubenswrapper[4715]: I1203 22:29:06.037469 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-jkd9h"] Dec 03 22:29:06 crc kubenswrapper[4715]: I1203 22:29:06.044604 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-jkd9h"] Dec 03 22:29:07 crc kubenswrapper[4715]: I1203 22:29:07.652794 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ac13b78-4213-452d-8619-9a2bc7ee5338" path="/var/lib/kubelet/pods/1ac13b78-4213-452d-8619-9a2bc7ee5338/volumes" Dec 03 22:29:09 crc kubenswrapper[4715]: I1203 22:29:09.038908 4715 generic.go:334] "Generic (PLEG): container finished" podID="db7da37a-b417-433c-86e9-a18446b4af27" containerID="2708ba6a22837a148100380497dc32ec7e01e3f7eb3d1d39e0ab627386f849f5" exitCode=0 Dec 03 22:29:09 crc kubenswrapper[4715]: I1203 22:29:09.039003 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" event={"ID":"db7da37a-b417-433c-86e9-a18446b4af27","Type":"ContainerDied","Data":"2708ba6a22837a148100380497dc32ec7e01e3f7eb3d1d39e0ab627386f849f5"} Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.534290 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.536975 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-inventory\") pod \"db7da37a-b417-433c-86e9-a18446b4af27\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.537096 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-ssh-key\") pod \"db7da37a-b417-433c-86e9-a18446b4af27\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.537126 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t495\" (UniqueName: \"kubernetes.io/projected/db7da37a-b417-433c-86e9-a18446b4af27-kube-api-access-4t495\") pod \"db7da37a-b417-433c-86e9-a18446b4af27\" (UID: \"db7da37a-b417-433c-86e9-a18446b4af27\") " Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.543853 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db7da37a-b417-433c-86e9-a18446b4af27-kube-api-access-4t495" (OuterVolumeSpecName: "kube-api-access-4t495") pod "db7da37a-b417-433c-86e9-a18446b4af27" (UID: "db7da37a-b417-433c-86e9-a18446b4af27"). InnerVolumeSpecName "kube-api-access-4t495". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.572251 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db7da37a-b417-433c-86e9-a18446b4af27" (UID: "db7da37a-b417-433c-86e9-a18446b4af27"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.575726 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-inventory" (OuterVolumeSpecName: "inventory") pod "db7da37a-b417-433c-86e9-a18446b4af27" (UID: "db7da37a-b417-433c-86e9-a18446b4af27"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.638335 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.638363 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db7da37a-b417-433c-86e9-a18446b4af27-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:10 crc kubenswrapper[4715]: I1203 22:29:10.638373 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t495\" (UniqueName: \"kubernetes.io/projected/db7da37a-b417-433c-86e9-a18446b4af27-kube-api-access-4t495\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.097591 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" event={"ID":"db7da37a-b417-433c-86e9-a18446b4af27","Type":"ContainerDied","Data":"be22066caf6f802d7e17ab0b7951c27e47a267434b8366299274f1b976e97a4a"} Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.097971 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be22066caf6f802d7e17ab0b7951c27e47a267434b8366299274f1b976e97a4a" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.098076 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.160000 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j"] Dec 03 22:29:11 crc kubenswrapper[4715]: E1203 22:29:11.160453 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db7da37a-b417-433c-86e9-a18446b4af27" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.160477 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="db7da37a-b417-433c-86e9-a18446b4af27" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.160717 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="db7da37a-b417-433c-86e9-a18446b4af27" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.161494 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.164182 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.164235 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.164489 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.164655 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.183881 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j"] Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.253544 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.253601 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rjk6\" (UniqueName: \"kubernetes.io/projected/3976028c-4d85-4518-9151-0e7c94638d58-kube-api-access-2rjk6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.253632 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.355410 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.355450 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rjk6\" (UniqueName: \"kubernetes.io/projected/3976028c-4d85-4518-9151-0e7c94638d58-kube-api-access-2rjk6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.355478 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.360541 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.360759 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.372767 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rjk6\" (UniqueName: \"kubernetes.io/projected/3976028c-4d85-4518-9151-0e7c94638d58-kube-api-access-2rjk6\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-pmd4j\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.488911 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:11 crc kubenswrapper[4715]: I1203 22:29:11.855425 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j"] Dec 03 22:29:11 crc kubenswrapper[4715]: W1203 22:29:11.859925 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3976028c_4d85_4518_9151_0e7c94638d58.slice/crio-1a08e600e2db9aff053cf2de8b5cc986d02de071f8b26d978f6a43bec026800d WatchSource:0}: Error finding container 1a08e600e2db9aff053cf2de8b5cc986d02de071f8b26d978f6a43bec026800d: Status 404 returned error can't find the container with id 1a08e600e2db9aff053cf2de8b5cc986d02de071f8b26d978f6a43bec026800d Dec 03 22:29:12 crc kubenswrapper[4715]: I1203 22:29:12.111051 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" event={"ID":"3976028c-4d85-4518-9151-0e7c94638d58","Type":"ContainerStarted","Data":"1a08e600e2db9aff053cf2de8b5cc986d02de071f8b26d978f6a43bec026800d"} Dec 03 22:29:13 crc kubenswrapper[4715]: I1203 22:29:13.123934 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" event={"ID":"3976028c-4d85-4518-9151-0e7c94638d58","Type":"ContainerStarted","Data":"604368ee4a1551fcc49741af88e4214115518dbc87590ffaac3c1822ae0e68d2"} Dec 03 22:29:13 crc kubenswrapper[4715]: I1203 22:29:13.145742 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" podStartSLOduration=1.6315744859999999 podStartE2EDuration="2.145722683s" podCreationTimestamp="2025-12-03 22:29:11 +0000 UTC" firstStartedPulling="2025-12-03 22:29:11.863555135 +0000 UTC m=+2848.606265740" lastFinishedPulling="2025-12-03 22:29:12.377703332 +0000 UTC m=+2849.120413937" observedRunningTime="2025-12-03 22:29:13.142897367 +0000 UTC m=+2849.885607972" watchObservedRunningTime="2025-12-03 22:29:13.145722683 +0000 UTC m=+2849.888433278" Dec 03 22:29:14 crc kubenswrapper[4715]: I1203 22:29:14.040903 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-svdtv"] Dec 03 22:29:14 crc kubenswrapper[4715]: I1203 22:29:14.050864 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-svdtv"] Dec 03 22:29:15 crc kubenswrapper[4715]: I1203 22:29:15.647885 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f05149da-e2c3-4bbb-9b7b-e51f6d022875" path="/var/lib/kubelet/pods/f05149da-e2c3-4bbb-9b7b-e51f6d022875/volumes" Dec 03 22:29:35 crc kubenswrapper[4715]: I1203 22:29:35.159613 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:29:35 crc kubenswrapper[4715]: I1203 22:29:35.160453 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:29:55 crc kubenswrapper[4715]: I1203 22:29:55.268360 4715 scope.go:117] "RemoveContainer" containerID="d38a7035297765d245716784ed4f49f1e33b03bdd2f0cd344fa42d5fddd3f62a" Dec 03 22:29:55 crc kubenswrapper[4715]: I1203 22:29:55.317835 4715 scope.go:117] "RemoveContainer" containerID="9260bbec95f69642bf1876790b1f2325d6377636275474c7052e74f51c546005" Dec 03 22:29:56 crc kubenswrapper[4715]: I1203 22:29:56.571852 4715 generic.go:334] "Generic (PLEG): container finished" podID="3976028c-4d85-4518-9151-0e7c94638d58" containerID="604368ee4a1551fcc49741af88e4214115518dbc87590ffaac3c1822ae0e68d2" exitCode=0 Dec 03 22:29:56 crc kubenswrapper[4715]: I1203 22:29:56.571904 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" event={"ID":"3976028c-4d85-4518-9151-0e7c94638d58","Type":"ContainerDied","Data":"604368ee4a1551fcc49741af88e4214115518dbc87590ffaac3c1822ae0e68d2"} Dec 03 22:29:57 crc kubenswrapper[4715]: I1203 22:29:57.045767 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-bhkbs"] Dec 03 22:29:57 crc kubenswrapper[4715]: I1203 22:29:57.055491 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-bhkbs"] Dec 03 22:29:57 crc kubenswrapper[4715]: I1203 22:29:57.649863 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="693acee8-7f5a-40ce-9b09-57c92bfd97c1" path="/var/lib/kubelet/pods/693acee8-7f5a-40ce-9b09-57c92bfd97c1/volumes" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.040793 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.048546 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rjk6\" (UniqueName: \"kubernetes.io/projected/3976028c-4d85-4518-9151-0e7c94638d58-kube-api-access-2rjk6\") pod \"3976028c-4d85-4518-9151-0e7c94638d58\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.048639 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-inventory\") pod \"3976028c-4d85-4518-9151-0e7c94638d58\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.048670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-ssh-key\") pod \"3976028c-4d85-4518-9151-0e7c94638d58\" (UID: \"3976028c-4d85-4518-9151-0e7c94638d58\") " Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.053959 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3976028c-4d85-4518-9151-0e7c94638d58-kube-api-access-2rjk6" (OuterVolumeSpecName: "kube-api-access-2rjk6") pod "3976028c-4d85-4518-9151-0e7c94638d58" (UID: "3976028c-4d85-4518-9151-0e7c94638d58"). InnerVolumeSpecName "kube-api-access-2rjk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.077534 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-inventory" (OuterVolumeSpecName: "inventory") pod "3976028c-4d85-4518-9151-0e7c94638d58" (UID: "3976028c-4d85-4518-9151-0e7c94638d58"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.086202 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3976028c-4d85-4518-9151-0e7c94638d58" (UID: "3976028c-4d85-4518-9151-0e7c94638d58"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.150028 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rjk6\" (UniqueName: \"kubernetes.io/projected/3976028c-4d85-4518-9151-0e7c94638d58-kube-api-access-2rjk6\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.150058 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.150068 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3976028c-4d85-4518-9151-0e7c94638d58-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.599186 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" event={"ID":"3976028c-4d85-4518-9151-0e7c94638d58","Type":"ContainerDied","Data":"1a08e600e2db9aff053cf2de8b5cc986d02de071f8b26d978f6a43bec026800d"} Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.599289 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a08e600e2db9aff053cf2de8b5cc986d02de071f8b26d978f6a43bec026800d" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.599299 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-pmd4j" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.721155 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz"] Dec 03 22:29:58 crc kubenswrapper[4715]: E1203 22:29:58.721741 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3976028c-4d85-4518-9151-0e7c94638d58" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.721768 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3976028c-4d85-4518-9151-0e7c94638d58" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.722025 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3976028c-4d85-4518-9151-0e7c94638d58" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.722904 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.725179 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.727856 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.732930 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.736156 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.750033 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz"] Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.761334 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.761530 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.761703 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwxqm\" (UniqueName: \"kubernetes.io/projected/b1a79612-747c-4bbd-be8c-f79a9da177db-kube-api-access-fwxqm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.863905 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwxqm\" (UniqueName: \"kubernetes.io/projected/b1a79612-747c-4bbd-be8c-f79a9da177db-kube-api-access-fwxqm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.864020 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.864143 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.868397 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.868808 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:58 crc kubenswrapper[4715]: I1203 22:29:58.889014 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwxqm\" (UniqueName: \"kubernetes.io/projected/b1a79612-747c-4bbd-be8c-f79a9da177db-kube-api-access-fwxqm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-9mszz\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:59 crc kubenswrapper[4715]: I1203 22:29:59.050996 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:29:59 crc kubenswrapper[4715]: I1203 22:29:59.604342 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz"] Dec 03 22:29:59 crc kubenswrapper[4715]: W1203 22:29:59.613759 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1a79612_747c_4bbd_be8c_f79a9da177db.slice/crio-089e0c15bea8d8199f88c20b62b7979d52524e80033a58b6d83397774e9ef121 WatchSource:0}: Error finding container 089e0c15bea8d8199f88c20b62b7979d52524e80033a58b6d83397774e9ef121: Status 404 returned error can't find the container with id 089e0c15bea8d8199f88c20b62b7979d52524e80033a58b6d83397774e9ef121 Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.135587 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr"] Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.137180 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.139991 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.141201 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.146450 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr"] Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.300044 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz4rr\" (UniqueName: \"kubernetes.io/projected/633f03e8-4172-4791-8e99-c7f31ae34dba-kube-api-access-kz4rr\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.300118 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/633f03e8-4172-4791-8e99-c7f31ae34dba-secret-volume\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.300210 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/633f03e8-4172-4791-8e99-c7f31ae34dba-config-volume\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.401742 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/633f03e8-4172-4791-8e99-c7f31ae34dba-config-volume\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.402225 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz4rr\" (UniqueName: \"kubernetes.io/projected/633f03e8-4172-4791-8e99-c7f31ae34dba-kube-api-access-kz4rr\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.402282 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/633f03e8-4172-4791-8e99-c7f31ae34dba-secret-volume\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.402981 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/633f03e8-4172-4791-8e99-c7f31ae34dba-config-volume\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.407261 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/633f03e8-4172-4791-8e99-c7f31ae34dba-secret-volume\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.425830 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz4rr\" (UniqueName: \"kubernetes.io/projected/633f03e8-4172-4791-8e99-c7f31ae34dba-kube-api-access-kz4rr\") pod \"collect-profiles-29413350-4h5tr\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.519037 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.621568 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" event={"ID":"b1a79612-747c-4bbd-be8c-f79a9da177db","Type":"ContainerStarted","Data":"9c325a010c53034727a89984a70c133fbc18eba089955a2580829e0411e82015"} Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.621642 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" event={"ID":"b1a79612-747c-4bbd-be8c-f79a9da177db","Type":"ContainerStarted","Data":"089e0c15bea8d8199f88c20b62b7979d52524e80033a58b6d83397774e9ef121"} Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.645948 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" podStartSLOduration=2.250440136 podStartE2EDuration="2.645922979s" podCreationTimestamp="2025-12-03 22:29:58 +0000 UTC" firstStartedPulling="2025-12-03 22:29:59.615553917 +0000 UTC m=+2896.358264512" lastFinishedPulling="2025-12-03 22:30:00.01103675 +0000 UTC m=+2896.753747355" observedRunningTime="2025-12-03 22:30:00.639400714 +0000 UTC m=+2897.382111309" watchObservedRunningTime="2025-12-03 22:30:00.645922979 +0000 UTC m=+2897.388633584" Dec 03 22:30:00 crc kubenswrapper[4715]: I1203 22:30:00.942044 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr"] Dec 03 22:30:00 crc kubenswrapper[4715]: W1203 22:30:00.942910 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod633f03e8_4172_4791_8e99_c7f31ae34dba.slice/crio-9f503ef8adc542c14909794c181971c9bb41c9c164282d80795672740e3a7f3e WatchSource:0}: Error finding container 9f503ef8adc542c14909794c181971c9bb41c9c164282d80795672740e3a7f3e: Status 404 returned error can't find the container with id 9f503ef8adc542c14909794c181971c9bb41c9c164282d80795672740e3a7f3e Dec 03 22:30:01 crc kubenswrapper[4715]: I1203 22:30:01.634745 4715 generic.go:334] "Generic (PLEG): container finished" podID="633f03e8-4172-4791-8e99-c7f31ae34dba" containerID="b9eeeb7de6354154d758d2eb5bd6db2cb6425df7fd02c47d4efa8e091f685a2c" exitCode=0 Dec 03 22:30:01 crc kubenswrapper[4715]: I1203 22:30:01.655206 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" event={"ID":"633f03e8-4172-4791-8e99-c7f31ae34dba","Type":"ContainerDied","Data":"b9eeeb7de6354154d758d2eb5bd6db2cb6425df7fd02c47d4efa8e091f685a2c"} Dec 03 22:30:01 crc kubenswrapper[4715]: I1203 22:30:01.655268 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" event={"ID":"633f03e8-4172-4791-8e99-c7f31ae34dba","Type":"ContainerStarted","Data":"9f503ef8adc542c14909794c181971c9bb41c9c164282d80795672740e3a7f3e"} Dec 03 22:30:02 crc kubenswrapper[4715]: I1203 22:30:02.983480 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.155128 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/633f03e8-4172-4791-8e99-c7f31ae34dba-secret-volume\") pod \"633f03e8-4172-4791-8e99-c7f31ae34dba\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.155460 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/633f03e8-4172-4791-8e99-c7f31ae34dba-config-volume\") pod \"633f03e8-4172-4791-8e99-c7f31ae34dba\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.155530 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz4rr\" (UniqueName: \"kubernetes.io/projected/633f03e8-4172-4791-8e99-c7f31ae34dba-kube-api-access-kz4rr\") pod \"633f03e8-4172-4791-8e99-c7f31ae34dba\" (UID: \"633f03e8-4172-4791-8e99-c7f31ae34dba\") " Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.156278 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/633f03e8-4172-4791-8e99-c7f31ae34dba-config-volume" (OuterVolumeSpecName: "config-volume") pod "633f03e8-4172-4791-8e99-c7f31ae34dba" (UID: "633f03e8-4172-4791-8e99-c7f31ae34dba"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.160973 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/633f03e8-4172-4791-8e99-c7f31ae34dba-kube-api-access-kz4rr" (OuterVolumeSpecName: "kube-api-access-kz4rr") pod "633f03e8-4172-4791-8e99-c7f31ae34dba" (UID: "633f03e8-4172-4791-8e99-c7f31ae34dba"). InnerVolumeSpecName "kube-api-access-kz4rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.161548 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/633f03e8-4172-4791-8e99-c7f31ae34dba-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "633f03e8-4172-4791-8e99-c7f31ae34dba" (UID: "633f03e8-4172-4791-8e99-c7f31ae34dba"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.257707 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz4rr\" (UniqueName: \"kubernetes.io/projected/633f03e8-4172-4791-8e99-c7f31ae34dba-kube-api-access-kz4rr\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.258047 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/633f03e8-4172-4791-8e99-c7f31ae34dba-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.258070 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/633f03e8-4172-4791-8e99-c7f31ae34dba-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.657368 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" event={"ID":"633f03e8-4172-4791-8e99-c7f31ae34dba","Type":"ContainerDied","Data":"9f503ef8adc542c14909794c181971c9bb41c9c164282d80795672740e3a7f3e"} Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.657404 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f503ef8adc542c14909794c181971c9bb41c9c164282d80795672740e3a7f3e" Dec 03 22:30:03 crc kubenswrapper[4715]: I1203 22:30:03.657460 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr" Dec 03 22:30:04 crc kubenswrapper[4715]: I1203 22:30:04.050185 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs"] Dec 03 22:30:04 crc kubenswrapper[4715]: I1203 22:30:04.057616 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413305-5xcgs"] Dec 03 22:30:05 crc kubenswrapper[4715]: I1203 22:30:05.159411 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:30:05 crc kubenswrapper[4715]: I1203 22:30:05.159459 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:30:05 crc kubenswrapper[4715]: I1203 22:30:05.646291 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55aa040e-0a60-41b2-add8-f2900df05cdb" path="/var/lib/kubelet/pods/55aa040e-0a60-41b2-add8-f2900df05cdb/volumes" Dec 03 22:30:17 crc kubenswrapper[4715]: I1203 22:30:17.896157 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zngv8"] Dec 03 22:30:17 crc kubenswrapper[4715]: E1203 22:30:17.899223 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="633f03e8-4172-4791-8e99-c7f31ae34dba" containerName="collect-profiles" Dec 03 22:30:17 crc kubenswrapper[4715]: I1203 22:30:17.899245 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="633f03e8-4172-4791-8e99-c7f31ae34dba" containerName="collect-profiles" Dec 03 22:30:17 crc kubenswrapper[4715]: I1203 22:30:17.899557 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="633f03e8-4172-4791-8e99-c7f31ae34dba" containerName="collect-profiles" Dec 03 22:30:17 crc kubenswrapper[4715]: I1203 22:30:17.901370 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:17 crc kubenswrapper[4715]: I1203 22:30:17.909933 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zngv8"] Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.072314 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-catalog-content\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.072419 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phjvv\" (UniqueName: \"kubernetes.io/projected/de2c5030-260b-4b4a-b664-eeec5082454b-kube-api-access-phjvv\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.072641 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-utilities\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.174917 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phjvv\" (UniqueName: \"kubernetes.io/projected/de2c5030-260b-4b4a-b664-eeec5082454b-kube-api-access-phjvv\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.175058 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-utilities\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.175105 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-catalog-content\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.175562 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-catalog-content\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.175626 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-utilities\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.196901 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phjvv\" (UniqueName: \"kubernetes.io/projected/de2c5030-260b-4b4a-b664-eeec5082454b-kube-api-access-phjvv\") pod \"redhat-operators-zngv8\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.220772 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.662963 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zngv8"] Dec 03 22:30:18 crc kubenswrapper[4715]: I1203 22:30:18.790959 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zngv8" event={"ID":"de2c5030-260b-4b4a-b664-eeec5082454b","Type":"ContainerStarted","Data":"e839c5f129a705f2a0c456f230dde61ac20782a937b8550daef6b01652f68a9d"} Dec 03 22:30:19 crc kubenswrapper[4715]: I1203 22:30:19.803466 4715 generic.go:334] "Generic (PLEG): container finished" podID="de2c5030-260b-4b4a-b664-eeec5082454b" containerID="c053f3799f1be1889caee16b90993481f26eca23583678277ffee33290d57852" exitCode=0 Dec 03 22:30:19 crc kubenswrapper[4715]: I1203 22:30:19.803534 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zngv8" event={"ID":"de2c5030-260b-4b4a-b664-eeec5082454b","Type":"ContainerDied","Data":"c053f3799f1be1889caee16b90993481f26eca23583678277ffee33290d57852"} Dec 03 22:30:20 crc kubenswrapper[4715]: I1203 22:30:20.816167 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zngv8" event={"ID":"de2c5030-260b-4b4a-b664-eeec5082454b","Type":"ContainerStarted","Data":"362c04a5b5f1d8efc4bdc545ee752439d0aeb6534f7abb8fec5e6bf2e52f5c58"} Dec 03 22:30:21 crc kubenswrapper[4715]: I1203 22:30:21.837855 4715 generic.go:334] "Generic (PLEG): container finished" podID="de2c5030-260b-4b4a-b664-eeec5082454b" containerID="362c04a5b5f1d8efc4bdc545ee752439d0aeb6534f7abb8fec5e6bf2e52f5c58" exitCode=0 Dec 03 22:30:21 crc kubenswrapper[4715]: I1203 22:30:21.837958 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zngv8" event={"ID":"de2c5030-260b-4b4a-b664-eeec5082454b","Type":"ContainerDied","Data":"362c04a5b5f1d8efc4bdc545ee752439d0aeb6534f7abb8fec5e6bf2e52f5c58"} Dec 03 22:30:24 crc kubenswrapper[4715]: I1203 22:30:24.868647 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zngv8" event={"ID":"de2c5030-260b-4b4a-b664-eeec5082454b","Type":"ContainerStarted","Data":"e36b567f689e85d92274b49056b93c6f2a17fa3a018f2fc87db7edfa0dcb6066"} Dec 03 22:30:24 crc kubenswrapper[4715]: I1203 22:30:24.889991 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zngv8" podStartSLOduration=3.525142031 podStartE2EDuration="7.889967361s" podCreationTimestamp="2025-12-03 22:30:17 +0000 UTC" firstStartedPulling="2025-12-03 22:30:19.807328928 +0000 UTC m=+2916.550039523" lastFinishedPulling="2025-12-03 22:30:24.172154248 +0000 UTC m=+2920.914864853" observedRunningTime="2025-12-03 22:30:24.885968471 +0000 UTC m=+2921.628679066" watchObservedRunningTime="2025-12-03 22:30:24.889967361 +0000 UTC m=+2921.632677966" Dec 03 22:30:28 crc kubenswrapper[4715]: I1203 22:30:28.221913 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:28 crc kubenswrapper[4715]: I1203 22:30:28.222324 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:29 crc kubenswrapper[4715]: I1203 22:30:29.282415 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zngv8" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="registry-server" probeResult="failure" output=< Dec 03 22:30:29 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 03 22:30:29 crc kubenswrapper[4715]: > Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.160419 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.161329 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.161402 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.162407 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.162484 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" gracePeriod=600 Dec 03 22:30:35 crc kubenswrapper[4715]: E1203 22:30:35.793351 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.974712 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" exitCode=0 Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.974757 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a"} Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.974861 4715 scope.go:117] "RemoveContainer" containerID="33ce01d1db2f2fd17a93394b524ae7fd388810ef368375808b76327aec02d590" Dec 03 22:30:35 crc kubenswrapper[4715]: I1203 22:30:35.975871 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:30:35 crc kubenswrapper[4715]: E1203 22:30:35.976426 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:30:38 crc kubenswrapper[4715]: I1203 22:30:38.287003 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:38 crc kubenswrapper[4715]: I1203 22:30:38.339107 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:38 crc kubenswrapper[4715]: I1203 22:30:38.531106 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zngv8"] Dec 03 22:30:40 crc kubenswrapper[4715]: I1203 22:30:40.032890 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zngv8" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="registry-server" containerID="cri-o://e36b567f689e85d92274b49056b93c6f2a17fa3a018f2fc87db7edfa0dcb6066" gracePeriod=2 Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.044609 4715 generic.go:334] "Generic (PLEG): container finished" podID="de2c5030-260b-4b4a-b664-eeec5082454b" containerID="e36b567f689e85d92274b49056b93c6f2a17fa3a018f2fc87db7edfa0dcb6066" exitCode=0 Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.044655 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zngv8" event={"ID":"de2c5030-260b-4b4a-b664-eeec5082454b","Type":"ContainerDied","Data":"e36b567f689e85d92274b49056b93c6f2a17fa3a018f2fc87db7edfa0dcb6066"} Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.631147 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.769485 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-catalog-content\") pod \"de2c5030-260b-4b4a-b664-eeec5082454b\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.769615 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phjvv\" (UniqueName: \"kubernetes.io/projected/de2c5030-260b-4b4a-b664-eeec5082454b-kube-api-access-phjvv\") pod \"de2c5030-260b-4b4a-b664-eeec5082454b\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.769753 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-utilities\") pod \"de2c5030-260b-4b4a-b664-eeec5082454b\" (UID: \"de2c5030-260b-4b4a-b664-eeec5082454b\") " Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.771160 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-utilities" (OuterVolumeSpecName: "utilities") pod "de2c5030-260b-4b4a-b664-eeec5082454b" (UID: "de2c5030-260b-4b4a-b664-eeec5082454b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.777659 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2c5030-260b-4b4a-b664-eeec5082454b-kube-api-access-phjvv" (OuterVolumeSpecName: "kube-api-access-phjvv") pod "de2c5030-260b-4b4a-b664-eeec5082454b" (UID: "de2c5030-260b-4b4a-b664-eeec5082454b"). InnerVolumeSpecName "kube-api-access-phjvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.872324 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phjvv\" (UniqueName: \"kubernetes.io/projected/de2c5030-260b-4b4a-b664-eeec5082454b-kube-api-access-phjvv\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.872359 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.874255 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de2c5030-260b-4b4a-b664-eeec5082454b" (UID: "de2c5030-260b-4b4a-b664-eeec5082454b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:30:41 crc kubenswrapper[4715]: I1203 22:30:41.973833 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de2c5030-260b-4b4a-b664-eeec5082454b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:42 crc kubenswrapper[4715]: I1203 22:30:42.061226 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zngv8" event={"ID":"de2c5030-260b-4b4a-b664-eeec5082454b","Type":"ContainerDied","Data":"e839c5f129a705f2a0c456f230dde61ac20782a937b8550daef6b01652f68a9d"} Dec 03 22:30:42 crc kubenswrapper[4715]: I1203 22:30:42.061271 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zngv8" Dec 03 22:30:42 crc kubenswrapper[4715]: I1203 22:30:42.061294 4715 scope.go:117] "RemoveContainer" containerID="e36b567f689e85d92274b49056b93c6f2a17fa3a018f2fc87db7edfa0dcb6066" Dec 03 22:30:42 crc kubenswrapper[4715]: I1203 22:30:42.096076 4715 scope.go:117] "RemoveContainer" containerID="362c04a5b5f1d8efc4bdc545ee752439d0aeb6534f7abb8fec5e6bf2e52f5c58" Dec 03 22:30:42 crc kubenswrapper[4715]: I1203 22:30:42.125373 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zngv8"] Dec 03 22:30:42 crc kubenswrapper[4715]: I1203 22:30:42.136459 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zngv8"] Dec 03 22:30:42 crc kubenswrapper[4715]: I1203 22:30:42.157244 4715 scope.go:117] "RemoveContainer" containerID="c053f3799f1be1889caee16b90993481f26eca23583678277ffee33290d57852" Dec 03 22:30:43 crc kubenswrapper[4715]: I1203 22:30:43.653349 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" path="/var/lib/kubelet/pods/de2c5030-260b-4b4a-b664-eeec5082454b/volumes" Dec 03 22:30:50 crc kubenswrapper[4715]: I1203 22:30:50.634208 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:30:50 crc kubenswrapper[4715]: E1203 22:30:50.635348 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:30:54 crc kubenswrapper[4715]: I1203 22:30:54.224166 4715 generic.go:334] "Generic (PLEG): container finished" podID="b1a79612-747c-4bbd-be8c-f79a9da177db" containerID="9c325a010c53034727a89984a70c133fbc18eba089955a2580829e0411e82015" exitCode=0 Dec 03 22:30:54 crc kubenswrapper[4715]: I1203 22:30:54.224299 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" event={"ID":"b1a79612-747c-4bbd-be8c-f79a9da177db","Type":"ContainerDied","Data":"9c325a010c53034727a89984a70c133fbc18eba089955a2580829e0411e82015"} Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.458654 4715 scope.go:117] "RemoveContainer" containerID="d31c7d87d9acbaaa4018c190ce98e56cff097021a072ff198f044636d13c0999" Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.494868 4715 scope.go:117] "RemoveContainer" containerID="bccc200445d5a2106f8a85baa68217d4e6b0dde973a823ea144fcb39d7835e7c" Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.619287 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.691255 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-ssh-key\") pod \"b1a79612-747c-4bbd-be8c-f79a9da177db\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.691371 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwxqm\" (UniqueName: \"kubernetes.io/projected/b1a79612-747c-4bbd-be8c-f79a9da177db-kube-api-access-fwxqm\") pod \"b1a79612-747c-4bbd-be8c-f79a9da177db\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.691418 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-inventory\") pod \"b1a79612-747c-4bbd-be8c-f79a9da177db\" (UID: \"b1a79612-747c-4bbd-be8c-f79a9da177db\") " Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.697217 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1a79612-747c-4bbd-be8c-f79a9da177db-kube-api-access-fwxqm" (OuterVolumeSpecName: "kube-api-access-fwxqm") pod "b1a79612-747c-4bbd-be8c-f79a9da177db" (UID: "b1a79612-747c-4bbd-be8c-f79a9da177db"). InnerVolumeSpecName "kube-api-access-fwxqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.717408 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b1a79612-747c-4bbd-be8c-f79a9da177db" (UID: "b1a79612-747c-4bbd-be8c-f79a9da177db"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.721015 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-inventory" (OuterVolumeSpecName: "inventory") pod "b1a79612-747c-4bbd-be8c-f79a9da177db" (UID: "b1a79612-747c-4bbd-be8c-f79a9da177db"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.795288 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.796245 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwxqm\" (UniqueName: \"kubernetes.io/projected/b1a79612-747c-4bbd-be8c-f79a9da177db-kube-api-access-fwxqm\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:55 crc kubenswrapper[4715]: I1203 22:30:55.796273 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1a79612-747c-4bbd-be8c-f79a9da177db-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.247922 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" event={"ID":"b1a79612-747c-4bbd-be8c-f79a9da177db","Type":"ContainerDied","Data":"089e0c15bea8d8199f88c20b62b7979d52524e80033a58b6d83397774e9ef121"} Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.247964 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="089e0c15bea8d8199f88c20b62b7979d52524e80033a58b6d83397774e9ef121" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.247989 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-9mszz" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.358698 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2jmsp"] Dec 03 22:30:56 crc kubenswrapper[4715]: E1203 22:30:56.359386 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a79612-747c-4bbd-be8c-f79a9da177db" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.359420 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a79612-747c-4bbd-be8c-f79a9da177db" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:30:56 crc kubenswrapper[4715]: E1203 22:30:56.359464 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="extract-content" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.359477 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="extract-content" Dec 03 22:30:56 crc kubenswrapper[4715]: E1203 22:30:56.359529 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="registry-server" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.359541 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="registry-server" Dec 03 22:30:56 crc kubenswrapper[4715]: E1203 22:30:56.359562 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="extract-utilities" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.359572 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="extract-utilities" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.359856 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a79612-747c-4bbd-be8c-f79a9da177db" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.359893 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2c5030-260b-4b4a-b664-eeec5082454b" containerName="registry-server" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.360913 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.365968 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.366093 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.366419 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.366728 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.372015 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2jmsp"] Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.510309 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.510451 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.510536 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks4bw\" (UniqueName: \"kubernetes.io/projected/2e622ea8-2426-4e53-8775-3880cce092b2-kube-api-access-ks4bw\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.613189 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.613373 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.613458 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks4bw\" (UniqueName: \"kubernetes.io/projected/2e622ea8-2426-4e53-8775-3880cce092b2-kube-api-access-ks4bw\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.618253 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.621007 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.643059 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks4bw\" (UniqueName: \"kubernetes.io/projected/2e622ea8-2426-4e53-8775-3880cce092b2-kube-api-access-ks4bw\") pod \"ssh-known-hosts-edpm-deployment-2jmsp\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:56 crc kubenswrapper[4715]: I1203 22:30:56.686678 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:30:57 crc kubenswrapper[4715]: I1203 22:30:57.273487 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2jmsp"] Dec 03 22:30:58 crc kubenswrapper[4715]: I1203 22:30:58.274757 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" event={"ID":"2e622ea8-2426-4e53-8775-3880cce092b2","Type":"ContainerStarted","Data":"1f82f0235beb7375ab5d2fb9d8df4b25f936324ca714a7f875fb988ff70dfa76"} Dec 03 22:30:58 crc kubenswrapper[4715]: I1203 22:30:58.275009 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" event={"ID":"2e622ea8-2426-4e53-8775-3880cce092b2","Type":"ContainerStarted","Data":"a6b9a8b97bdefdbec8cc0d006dc3628304bebb2906dc77361f89f818c30ecd59"} Dec 03 22:30:58 crc kubenswrapper[4715]: I1203 22:30:58.291948 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" podStartSLOduration=1.7519978630000002 podStartE2EDuration="2.291934989s" podCreationTimestamp="2025-12-03 22:30:56 +0000 UTC" firstStartedPulling="2025-12-03 22:30:57.282520113 +0000 UTC m=+2954.025230718" lastFinishedPulling="2025-12-03 22:30:57.822457249 +0000 UTC m=+2954.565167844" observedRunningTime="2025-12-03 22:30:58.290745016 +0000 UTC m=+2955.033455611" watchObservedRunningTime="2025-12-03 22:30:58.291934989 +0000 UTC m=+2955.034645584" Dec 03 22:31:02 crc kubenswrapper[4715]: I1203 22:31:02.634564 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:31:02 crc kubenswrapper[4715]: E1203 22:31:02.635479 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:31:05 crc kubenswrapper[4715]: I1203 22:31:05.368219 4715 generic.go:334] "Generic (PLEG): container finished" podID="2e622ea8-2426-4e53-8775-3880cce092b2" containerID="1f82f0235beb7375ab5d2fb9d8df4b25f936324ca714a7f875fb988ff70dfa76" exitCode=0 Dec 03 22:31:05 crc kubenswrapper[4715]: I1203 22:31:05.371065 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" event={"ID":"2e622ea8-2426-4e53-8775-3880cce092b2","Type":"ContainerDied","Data":"1f82f0235beb7375ab5d2fb9d8df4b25f936324ca714a7f875fb988ff70dfa76"} Dec 03 22:31:06 crc kubenswrapper[4715]: I1203 22:31:06.834061 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:31:06 crc kubenswrapper[4715]: I1203 22:31:06.954522 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-inventory-0\") pod \"2e622ea8-2426-4e53-8775-3880cce092b2\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " Dec 03 22:31:06 crc kubenswrapper[4715]: I1203 22:31:06.954682 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-ssh-key-openstack-edpm-ipam\") pod \"2e622ea8-2426-4e53-8775-3880cce092b2\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " Dec 03 22:31:06 crc kubenswrapper[4715]: I1203 22:31:06.954779 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks4bw\" (UniqueName: \"kubernetes.io/projected/2e622ea8-2426-4e53-8775-3880cce092b2-kube-api-access-ks4bw\") pod \"2e622ea8-2426-4e53-8775-3880cce092b2\" (UID: \"2e622ea8-2426-4e53-8775-3880cce092b2\") " Dec 03 22:31:06 crc kubenswrapper[4715]: I1203 22:31:06.960542 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e622ea8-2426-4e53-8775-3880cce092b2-kube-api-access-ks4bw" (OuterVolumeSpecName: "kube-api-access-ks4bw") pod "2e622ea8-2426-4e53-8775-3880cce092b2" (UID: "2e622ea8-2426-4e53-8775-3880cce092b2"). InnerVolumeSpecName "kube-api-access-ks4bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:31:06 crc kubenswrapper[4715]: I1203 22:31:06.983393 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2e622ea8-2426-4e53-8775-3880cce092b2" (UID: "2e622ea8-2426-4e53-8775-3880cce092b2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:31:06 crc kubenswrapper[4715]: I1203 22:31:06.985353 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "2e622ea8-2426-4e53-8775-3880cce092b2" (UID: "2e622ea8-2426-4e53-8775-3880cce092b2"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.057393 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ks4bw\" (UniqueName: \"kubernetes.io/projected/2e622ea8-2426-4e53-8775-3880cce092b2-kube-api-access-ks4bw\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.057454 4715 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.057475 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2e622ea8-2426-4e53-8775-3880cce092b2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.389371 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" event={"ID":"2e622ea8-2426-4e53-8775-3880cce092b2","Type":"ContainerDied","Data":"a6b9a8b97bdefdbec8cc0d006dc3628304bebb2906dc77361f89f818c30ecd59"} Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.389418 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6b9a8b97bdefdbec8cc0d006dc3628304bebb2906dc77361f89f818c30ecd59" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.389512 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2jmsp" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.564192 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c"] Dec 03 22:31:07 crc kubenswrapper[4715]: E1203 22:31:07.564905 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e622ea8-2426-4e53-8775-3880cce092b2" containerName="ssh-known-hosts-edpm-deployment" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.564921 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e622ea8-2426-4e53-8775-3880cce092b2" containerName="ssh-known-hosts-edpm-deployment" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.565098 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e622ea8-2426-4e53-8775-3880cce092b2" containerName="ssh-known-hosts-edpm-deployment" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.565761 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.568153 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.569030 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.569054 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.570264 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.573959 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c"] Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.667958 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.668036 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.668127 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq7qp\" (UniqueName: \"kubernetes.io/projected/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-kube-api-access-qq7qp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.769883 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.769940 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.770000 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq7qp\" (UniqueName: \"kubernetes.io/projected/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-kube-api-access-qq7qp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.773815 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.774298 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.789608 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq7qp\" (UniqueName: \"kubernetes.io/projected/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-kube-api-access-qq7qp\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-9hf4c\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:07 crc kubenswrapper[4715]: I1203 22:31:07.890978 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:08 crc kubenswrapper[4715]: I1203 22:31:08.444259 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c"] Dec 03 22:31:08 crc kubenswrapper[4715]: W1203 22:31:08.445561 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod849e6369_17eb_4b31_a5cc_f44dc07c6e1a.slice/crio-50e8d27d36ce1c5f8fa0f14aa1e0a38c551980eebc1e9f55047c433523c0d093 WatchSource:0}: Error finding container 50e8d27d36ce1c5f8fa0f14aa1e0a38c551980eebc1e9f55047c433523c0d093: Status 404 returned error can't find the container with id 50e8d27d36ce1c5f8fa0f14aa1e0a38c551980eebc1e9f55047c433523c0d093 Dec 03 22:31:09 crc kubenswrapper[4715]: I1203 22:31:09.410708 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" event={"ID":"849e6369-17eb-4b31-a5cc-f44dc07c6e1a","Type":"ContainerStarted","Data":"1f214f0af37ab5c44f54067a1a40755224658e14767c57ff8ffcb0a27fec1849"} Dec 03 22:31:09 crc kubenswrapper[4715]: I1203 22:31:09.411068 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" event={"ID":"849e6369-17eb-4b31-a5cc-f44dc07c6e1a","Type":"ContainerStarted","Data":"50e8d27d36ce1c5f8fa0f14aa1e0a38c551980eebc1e9f55047c433523c0d093"} Dec 03 22:31:13 crc kubenswrapper[4715]: I1203 22:31:13.642183 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:31:13 crc kubenswrapper[4715]: E1203 22:31:13.643344 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:31:17 crc kubenswrapper[4715]: I1203 22:31:17.506318 4715 generic.go:334] "Generic (PLEG): container finished" podID="849e6369-17eb-4b31-a5cc-f44dc07c6e1a" containerID="1f214f0af37ab5c44f54067a1a40755224658e14767c57ff8ffcb0a27fec1849" exitCode=0 Dec 03 22:31:17 crc kubenswrapper[4715]: I1203 22:31:17.506425 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" event={"ID":"849e6369-17eb-4b31-a5cc-f44dc07c6e1a","Type":"ContainerDied","Data":"1f214f0af37ab5c44f54067a1a40755224658e14767c57ff8ffcb0a27fec1849"} Dec 03 22:31:18 crc kubenswrapper[4715]: I1203 22:31:18.917596 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:18 crc kubenswrapper[4715]: I1203 22:31:18.987875 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq7qp\" (UniqueName: \"kubernetes.io/projected/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-kube-api-access-qq7qp\") pod \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " Dec 03 22:31:18 crc kubenswrapper[4715]: I1203 22:31:18.987941 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-inventory\") pod \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " Dec 03 22:31:18 crc kubenswrapper[4715]: I1203 22:31:18.987988 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-ssh-key\") pod \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\" (UID: \"849e6369-17eb-4b31-a5cc-f44dc07c6e1a\") " Dec 03 22:31:18 crc kubenswrapper[4715]: I1203 22:31:18.994927 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-kube-api-access-qq7qp" (OuterVolumeSpecName: "kube-api-access-qq7qp") pod "849e6369-17eb-4b31-a5cc-f44dc07c6e1a" (UID: "849e6369-17eb-4b31-a5cc-f44dc07c6e1a"). InnerVolumeSpecName "kube-api-access-qq7qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.015573 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "849e6369-17eb-4b31-a5cc-f44dc07c6e1a" (UID: "849e6369-17eb-4b31-a5cc-f44dc07c6e1a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.024686 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-inventory" (OuterVolumeSpecName: "inventory") pod "849e6369-17eb-4b31-a5cc-f44dc07c6e1a" (UID: "849e6369-17eb-4b31-a5cc-f44dc07c6e1a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.090295 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq7qp\" (UniqueName: \"kubernetes.io/projected/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-kube-api-access-qq7qp\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.090334 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.090342 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/849e6369-17eb-4b31-a5cc-f44dc07c6e1a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.522477 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" event={"ID":"849e6369-17eb-4b31-a5cc-f44dc07c6e1a","Type":"ContainerDied","Data":"50e8d27d36ce1c5f8fa0f14aa1e0a38c551980eebc1e9f55047c433523c0d093"} Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.522619 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50e8d27d36ce1c5f8fa0f14aa1e0a38c551980eebc1e9f55047c433523c0d093" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.522558 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-9hf4c" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.607802 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74"] Dec 03 22:31:19 crc kubenswrapper[4715]: E1203 22:31:19.608334 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="849e6369-17eb-4b31-a5cc-f44dc07c6e1a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.608356 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="849e6369-17eb-4b31-a5cc-f44dc07c6e1a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.608625 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="849e6369-17eb-4b31-a5cc-f44dc07c6e1a" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.609427 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.632422 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.632708 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.633209 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.635143 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.654899 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74"] Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.703270 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.703395 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kzsk\" (UniqueName: \"kubernetes.io/projected/7790e81c-da01-4013-81ad-32bd29bba24f-kube-api-access-7kzsk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.703479 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.805066 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.805193 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kzsk\" (UniqueName: \"kubernetes.io/projected/7790e81c-da01-4013-81ad-32bd29bba24f-kube-api-access-7kzsk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.805263 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.808842 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.809679 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.825054 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kzsk\" (UniqueName: \"kubernetes.io/projected/7790e81c-da01-4013-81ad-32bd29bba24f-kube-api-access-7kzsk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:19 crc kubenswrapper[4715]: I1203 22:31:19.952424 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:20 crc kubenswrapper[4715]: I1203 22:31:20.492734 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74"] Dec 03 22:31:20 crc kubenswrapper[4715]: W1203 22:31:20.495536 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7790e81c_da01_4013_81ad_32bd29bba24f.slice/crio-f41650ef2a5cbab2196cfb900e03ff61a12c3bf405c6750b4f04a858c7902bc7 WatchSource:0}: Error finding container f41650ef2a5cbab2196cfb900e03ff61a12c3bf405c6750b4f04a858c7902bc7: Status 404 returned error can't find the container with id f41650ef2a5cbab2196cfb900e03ff61a12c3bf405c6750b4f04a858c7902bc7 Dec 03 22:31:20 crc kubenswrapper[4715]: I1203 22:31:20.532728 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" event={"ID":"7790e81c-da01-4013-81ad-32bd29bba24f","Type":"ContainerStarted","Data":"f41650ef2a5cbab2196cfb900e03ff61a12c3bf405c6750b4f04a858c7902bc7"} Dec 03 22:31:21 crc kubenswrapper[4715]: I1203 22:31:21.544755 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" event={"ID":"7790e81c-da01-4013-81ad-32bd29bba24f","Type":"ContainerStarted","Data":"9543693d88ad155efb033dcc82cb59511fa873096512bb59ab500cf9bbc1b5fc"} Dec 03 22:31:21 crc kubenswrapper[4715]: I1203 22:31:21.567609 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" podStartSLOduration=2.079451809 podStartE2EDuration="2.567584462s" podCreationTimestamp="2025-12-03 22:31:19 +0000 UTC" firstStartedPulling="2025-12-03 22:31:20.49864664 +0000 UTC m=+2977.241357245" lastFinishedPulling="2025-12-03 22:31:20.986779293 +0000 UTC m=+2977.729489898" observedRunningTime="2025-12-03 22:31:21.563353625 +0000 UTC m=+2978.306064230" watchObservedRunningTime="2025-12-03 22:31:21.567584462 +0000 UTC m=+2978.310295077" Dec 03 22:31:25 crc kubenswrapper[4715]: I1203 22:31:25.634641 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:31:25 crc kubenswrapper[4715]: E1203 22:31:25.635371 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:31:31 crc kubenswrapper[4715]: I1203 22:31:31.648276 4715 generic.go:334] "Generic (PLEG): container finished" podID="7790e81c-da01-4013-81ad-32bd29bba24f" containerID="9543693d88ad155efb033dcc82cb59511fa873096512bb59ab500cf9bbc1b5fc" exitCode=0 Dec 03 22:31:31 crc kubenswrapper[4715]: I1203 22:31:31.653273 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" event={"ID":"7790e81c-da01-4013-81ad-32bd29bba24f","Type":"ContainerDied","Data":"9543693d88ad155efb033dcc82cb59511fa873096512bb59ab500cf9bbc1b5fc"} Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.043217 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.192255 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kzsk\" (UniqueName: \"kubernetes.io/projected/7790e81c-da01-4013-81ad-32bd29bba24f-kube-api-access-7kzsk\") pod \"7790e81c-da01-4013-81ad-32bd29bba24f\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.192301 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-ssh-key\") pod \"7790e81c-da01-4013-81ad-32bd29bba24f\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.192609 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-inventory\") pod \"7790e81c-da01-4013-81ad-32bd29bba24f\" (UID: \"7790e81c-da01-4013-81ad-32bd29bba24f\") " Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.197770 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7790e81c-da01-4013-81ad-32bd29bba24f-kube-api-access-7kzsk" (OuterVolumeSpecName: "kube-api-access-7kzsk") pod "7790e81c-da01-4013-81ad-32bd29bba24f" (UID: "7790e81c-da01-4013-81ad-32bd29bba24f"). InnerVolumeSpecName "kube-api-access-7kzsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.219008 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7790e81c-da01-4013-81ad-32bd29bba24f" (UID: "7790e81c-da01-4013-81ad-32bd29bba24f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.231274 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-inventory" (OuterVolumeSpecName: "inventory") pod "7790e81c-da01-4013-81ad-32bd29bba24f" (UID: "7790e81c-da01-4013-81ad-32bd29bba24f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.294897 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kzsk\" (UniqueName: \"kubernetes.io/projected/7790e81c-da01-4013-81ad-32bd29bba24f-kube-api-access-7kzsk\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.294928 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.294942 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7790e81c-da01-4013-81ad-32bd29bba24f-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.666813 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" event={"ID":"7790e81c-da01-4013-81ad-32bd29bba24f","Type":"ContainerDied","Data":"f41650ef2a5cbab2196cfb900e03ff61a12c3bf405c6750b4f04a858c7902bc7"} Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.666848 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f41650ef2a5cbab2196cfb900e03ff61a12c3bf405c6750b4f04a858c7902bc7" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.666900 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.761042 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm"] Dec 03 22:31:33 crc kubenswrapper[4715]: E1203 22:31:33.761698 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7790e81c-da01-4013-81ad-32bd29bba24f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.761731 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="7790e81c-da01-4013-81ad-32bd29bba24f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.762044 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="7790e81c-da01-4013-81ad-32bd29bba24f" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.763071 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.765836 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.766676 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.766754 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.766894 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.767203 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.767244 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.767370 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.767561 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.778217 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm"] Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917144 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917186 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s5hl\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-kube-api-access-2s5hl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917214 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917357 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917484 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917587 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917632 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917706 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917738 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917816 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.917886 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.918000 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.918065 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:33 crc kubenswrapper[4715]: I1203 22:31:33.918181 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.020380 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.020748 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s5hl\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-kube-api-access-2s5hl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.020775 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.020814 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.020892 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.020932 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.020956 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.020993 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.021016 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.021059 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.021096 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.021152 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.021198 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.021262 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.027404 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.028062 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.029129 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.029394 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.029484 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.029752 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.030023 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.030680 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.031035 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.031901 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.031998 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.032288 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.032788 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.046314 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s5hl\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-kube-api-access-2s5hl\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-c27sm\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.079075 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.615535 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm"] Dec 03 22:31:34 crc kubenswrapper[4715]: I1203 22:31:34.677174 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" event={"ID":"c58da0a9-c5be-4018-a322-7ba8ef960d52","Type":"ContainerStarted","Data":"e1a187f4523c08f7ad673cf839ca0010bca9a31cef523ff2a605a40cac9cf95a"} Dec 03 22:31:35 crc kubenswrapper[4715]: I1203 22:31:35.687468 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" event={"ID":"c58da0a9-c5be-4018-a322-7ba8ef960d52","Type":"ContainerStarted","Data":"c3bf7811bfcd29e04fa8af1274c492306512232d5ea0af0645db1d1031eada03"} Dec 03 22:31:35 crc kubenswrapper[4715]: I1203 22:31:35.718479 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" podStartSLOduration=2.205215493 podStartE2EDuration="2.718454615s" podCreationTimestamp="2025-12-03 22:31:33 +0000 UTC" firstStartedPulling="2025-12-03 22:31:34.621779222 +0000 UTC m=+2991.364489817" lastFinishedPulling="2025-12-03 22:31:35.135018344 +0000 UTC m=+2991.877728939" observedRunningTime="2025-12-03 22:31:35.705798328 +0000 UTC m=+2992.448508953" watchObservedRunningTime="2025-12-03 22:31:35.718454615 +0000 UTC m=+2992.461165220" Dec 03 22:31:36 crc kubenswrapper[4715]: I1203 22:31:36.635601 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:31:36 crc kubenswrapper[4715]: E1203 22:31:36.636529 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:31:51 crc kubenswrapper[4715]: I1203 22:31:51.635194 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:31:51 crc kubenswrapper[4715]: E1203 22:31:51.636104 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:32:03 crc kubenswrapper[4715]: I1203 22:32:03.644598 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:32:03 crc kubenswrapper[4715]: E1203 22:32:03.645371 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:32:15 crc kubenswrapper[4715]: I1203 22:32:15.470645 4715 generic.go:334] "Generic (PLEG): container finished" podID="c58da0a9-c5be-4018-a322-7ba8ef960d52" containerID="c3bf7811bfcd29e04fa8af1274c492306512232d5ea0af0645db1d1031eada03" exitCode=0 Dec 03 22:32:15 crc kubenswrapper[4715]: I1203 22:32:15.470713 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" event={"ID":"c58da0a9-c5be-4018-a322-7ba8ef960d52","Type":"ContainerDied","Data":"c3bf7811bfcd29e04fa8af1274c492306512232d5ea0af0645db1d1031eada03"} Dec 03 22:32:16 crc kubenswrapper[4715]: I1203 22:32:16.959819 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129127 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ssh-key\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129378 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-nova-combined-ca-bundle\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129459 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-ovn-default-certs-0\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129564 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ovn-combined-ca-bundle\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129600 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-inventory\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129670 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129703 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2s5hl\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-kube-api-access-2s5hl\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129753 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-bootstrap-combined-ca-bundle\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.129809 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-telemetry-combined-ca-bundle\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.130380 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-libvirt-combined-ca-bundle\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.130454 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.130482 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-repo-setup-combined-ca-bundle\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.130537 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-neutron-metadata-combined-ca-bundle\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.130606 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"c58da0a9-c5be-4018-a322-7ba8ef960d52\" (UID: \"c58da0a9-c5be-4018-a322-7ba8ef960d52\") " Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.136273 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.136343 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.137886 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.138193 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-kube-api-access-2s5hl" (OuterVolumeSpecName: "kube-api-access-2s5hl") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "kube-api-access-2s5hl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.138447 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.138715 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.138729 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.138762 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.138817 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.139176 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.140425 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.141779 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.162702 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-inventory" (OuterVolumeSpecName: "inventory") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.180958 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c58da0a9-c5be-4018-a322-7ba8ef960d52" (UID: "c58da0a9-c5be-4018-a322-7ba8ef960d52"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233832 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233881 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2s5hl\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-kube-api-access-2s5hl\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233895 4715 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233909 4715 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233924 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233937 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233949 4715 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233961 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233973 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233984 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.233996 4715 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.234008 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c58da0a9-c5be-4018-a322-7ba8ef960d52-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.234019 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.234030 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58da0a9-c5be-4018-a322-7ba8ef960d52-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.491592 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" event={"ID":"c58da0a9-c5be-4018-a322-7ba8ef960d52","Type":"ContainerDied","Data":"e1a187f4523c08f7ad673cf839ca0010bca9a31cef523ff2a605a40cac9cf95a"} Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.491638 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1a187f4523c08f7ad673cf839ca0010bca9a31cef523ff2a605a40cac9cf95a" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.491653 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-c27sm" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.586340 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb"] Dec 03 22:32:17 crc kubenswrapper[4715]: E1203 22:32:17.586856 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58da0a9-c5be-4018-a322-7ba8ef960d52" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.586884 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58da0a9-c5be-4018-a322-7ba8ef960d52" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.587132 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c58da0a9-c5be-4018-a322-7ba8ef960d52" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.587866 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.593052 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.599973 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.600225 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.600328 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.600383 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.607482 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb"] Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.746349 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.746435 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.746458 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.747167 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkdft\" (UniqueName: \"kubernetes.io/projected/d0c3bcdb-424a-4da7-9917-7a7853067044-kube-api-access-lkdft\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.747260 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d0c3bcdb-424a-4da7-9917-7a7853067044-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.849395 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkdft\" (UniqueName: \"kubernetes.io/projected/d0c3bcdb-424a-4da7-9917-7a7853067044-kube-api-access-lkdft\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.849748 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d0c3bcdb-424a-4da7-9917-7a7853067044-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.849827 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.849881 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.849905 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.851999 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d0c3bcdb-424a-4da7-9917-7a7853067044-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.855573 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.856202 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.856363 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.875721 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkdft\" (UniqueName: \"kubernetes.io/projected/d0c3bcdb-424a-4da7-9917-7a7853067044-kube-api-access-lkdft\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-68qjb\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:17 crc kubenswrapper[4715]: I1203 22:32:17.902523 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:32:18 crc kubenswrapper[4715]: I1203 22:32:18.417357 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb"] Dec 03 22:32:18 crc kubenswrapper[4715]: I1203 22:32:18.500818 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" event={"ID":"d0c3bcdb-424a-4da7-9917-7a7853067044","Type":"ContainerStarted","Data":"fa6532e0a74ce41bc9f8779a0c7fd32539632b2de768dad063f013c8dc83f82b"} Dec 03 22:32:18 crc kubenswrapper[4715]: I1203 22:32:18.634187 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:32:18 crc kubenswrapper[4715]: E1203 22:32:18.634485 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:32:19 crc kubenswrapper[4715]: I1203 22:32:19.510946 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" event={"ID":"d0c3bcdb-424a-4da7-9917-7a7853067044","Type":"ContainerStarted","Data":"2259347289d9e42f35b2f16d5ae5dc31a0e8eb8b2bde44c3338e60cd5181806b"} Dec 03 22:32:19 crc kubenswrapper[4715]: I1203 22:32:19.532088 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" podStartSLOduration=2.076207912 podStartE2EDuration="2.532069998s" podCreationTimestamp="2025-12-03 22:32:17 +0000 UTC" firstStartedPulling="2025-12-03 22:32:18.4235875 +0000 UTC m=+3035.166298095" lastFinishedPulling="2025-12-03 22:32:18.879449586 +0000 UTC m=+3035.622160181" observedRunningTime="2025-12-03 22:32:19.528427747 +0000 UTC m=+3036.271138342" watchObservedRunningTime="2025-12-03 22:32:19.532069998 +0000 UTC m=+3036.274780593" Dec 03 22:32:31 crc kubenswrapper[4715]: I1203 22:32:31.634234 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:32:31 crc kubenswrapper[4715]: E1203 22:32:31.634981 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:32:43 crc kubenswrapper[4715]: I1203 22:32:43.643654 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:32:43 crc kubenswrapper[4715]: E1203 22:32:43.644513 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:32:57 crc kubenswrapper[4715]: I1203 22:32:57.635108 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:32:57 crc kubenswrapper[4715]: E1203 22:32:57.636811 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:33:12 crc kubenswrapper[4715]: I1203 22:33:12.634824 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:33:12 crc kubenswrapper[4715]: E1203 22:33:12.635697 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:33:23 crc kubenswrapper[4715]: I1203 22:33:23.643126 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:33:23 crc kubenswrapper[4715]: E1203 22:33:23.643977 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:33:24 crc kubenswrapper[4715]: I1203 22:33:24.162892 4715 generic.go:334] "Generic (PLEG): container finished" podID="d0c3bcdb-424a-4da7-9917-7a7853067044" containerID="2259347289d9e42f35b2f16d5ae5dc31a0e8eb8b2bde44c3338e60cd5181806b" exitCode=0 Dec 03 22:33:24 crc kubenswrapper[4715]: I1203 22:33:24.163007 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" event={"ID":"d0c3bcdb-424a-4da7-9917-7a7853067044","Type":"ContainerDied","Data":"2259347289d9e42f35b2f16d5ae5dc31a0e8eb8b2bde44c3338e60cd5181806b"} Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.647742 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.762918 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ovn-combined-ca-bundle\") pod \"d0c3bcdb-424a-4da7-9917-7a7853067044\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.762978 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkdft\" (UniqueName: \"kubernetes.io/projected/d0c3bcdb-424a-4da7-9917-7a7853067044-kube-api-access-lkdft\") pod \"d0c3bcdb-424a-4da7-9917-7a7853067044\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.763102 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ssh-key\") pod \"d0c3bcdb-424a-4da7-9917-7a7853067044\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.763172 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d0c3bcdb-424a-4da7-9917-7a7853067044-ovncontroller-config-0\") pod \"d0c3bcdb-424a-4da7-9917-7a7853067044\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.763221 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-inventory\") pod \"d0c3bcdb-424a-4da7-9917-7a7853067044\" (UID: \"d0c3bcdb-424a-4da7-9917-7a7853067044\") " Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.769575 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d0c3bcdb-424a-4da7-9917-7a7853067044" (UID: "d0c3bcdb-424a-4da7-9917-7a7853067044"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.770121 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0c3bcdb-424a-4da7-9917-7a7853067044-kube-api-access-lkdft" (OuterVolumeSpecName: "kube-api-access-lkdft") pod "d0c3bcdb-424a-4da7-9917-7a7853067044" (UID: "d0c3bcdb-424a-4da7-9917-7a7853067044"). InnerVolumeSpecName "kube-api-access-lkdft". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.791414 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0c3bcdb-424a-4da7-9917-7a7853067044-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "d0c3bcdb-424a-4da7-9917-7a7853067044" (UID: "d0c3bcdb-424a-4da7-9917-7a7853067044"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.793859 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d0c3bcdb-424a-4da7-9917-7a7853067044" (UID: "d0c3bcdb-424a-4da7-9917-7a7853067044"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.805427 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-inventory" (OuterVolumeSpecName: "inventory") pod "d0c3bcdb-424a-4da7-9917-7a7853067044" (UID: "d0c3bcdb-424a-4da7-9917-7a7853067044"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.867987 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.868052 4715 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d0c3bcdb-424a-4da7-9917-7a7853067044-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.868071 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.868086 4715 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0c3bcdb-424a-4da7-9917-7a7853067044-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:33:25 crc kubenswrapper[4715]: I1203 22:33:25.868100 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkdft\" (UniqueName: \"kubernetes.io/projected/d0c3bcdb-424a-4da7-9917-7a7853067044-kube-api-access-lkdft\") on node \"crc\" DevicePath \"\"" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.183674 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" event={"ID":"d0c3bcdb-424a-4da7-9917-7a7853067044","Type":"ContainerDied","Data":"fa6532e0a74ce41bc9f8779a0c7fd32539632b2de768dad063f013c8dc83f82b"} Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.183713 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa6532e0a74ce41bc9f8779a0c7fd32539632b2de768dad063f013c8dc83f82b" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.183780 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-68qjb" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.380167 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d"] Dec 03 22:33:26 crc kubenswrapper[4715]: E1203 22:33:26.380662 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0c3bcdb-424a-4da7-9917-7a7853067044" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.380685 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0c3bcdb-424a-4da7-9917-7a7853067044" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.380934 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0c3bcdb-424a-4da7-9917-7a7853067044" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.381820 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.395723 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d"] Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.396480 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.396758 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.396949 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.397130 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.397761 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.397985 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.481267 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.481330 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.481382 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl7lz\" (UniqueName: \"kubernetes.io/projected/c3439842-caec-430b-a91e-0b03443eff15-kube-api-access-cl7lz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.481403 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.481539 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.481572 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.582939 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.583311 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl7lz\" (UniqueName: \"kubernetes.io/projected/c3439842-caec-430b-a91e-0b03443eff15-kube-api-access-cl7lz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.583349 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.583488 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.583541 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.583603 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.587697 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.588034 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.588550 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.589073 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.589490 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.598920 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl7lz\" (UniqueName: \"kubernetes.io/projected/c3439842-caec-430b-a91e-0b03443eff15-kube-api-access-cl7lz\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:26 crc kubenswrapper[4715]: I1203 22:33:26.743247 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:33:27 crc kubenswrapper[4715]: I1203 22:33:27.286772 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d"] Dec 03 22:33:27 crc kubenswrapper[4715]: W1203 22:33:27.288050 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3439842_caec_430b_a91e_0b03443eff15.slice/crio-1e0b2157d11051af96d5a2995c7191e84d8777145ece8a54ad5e2c8165d50996 WatchSource:0}: Error finding container 1e0b2157d11051af96d5a2995c7191e84d8777145ece8a54ad5e2c8165d50996: Status 404 returned error can't find the container with id 1e0b2157d11051af96d5a2995c7191e84d8777145ece8a54ad5e2c8165d50996 Dec 03 22:33:27 crc kubenswrapper[4715]: I1203 22:33:27.291768 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:33:28 crc kubenswrapper[4715]: I1203 22:33:28.201548 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" event={"ID":"c3439842-caec-430b-a91e-0b03443eff15","Type":"ContainerStarted","Data":"1a4479a4e3020a484bd66101c82b779f4b8e606e932f2caa0ffe54fd099930db"} Dec 03 22:33:28 crc kubenswrapper[4715]: I1203 22:33:28.201978 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" event={"ID":"c3439842-caec-430b-a91e-0b03443eff15","Type":"ContainerStarted","Data":"1e0b2157d11051af96d5a2995c7191e84d8777145ece8a54ad5e2c8165d50996"} Dec 03 22:33:29 crc kubenswrapper[4715]: I1203 22:33:29.234600 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" podStartSLOduration=2.70488338 podStartE2EDuration="3.234573154s" podCreationTimestamp="2025-12-03 22:33:26 +0000 UTC" firstStartedPulling="2025-12-03 22:33:27.291558067 +0000 UTC m=+3104.034268662" lastFinishedPulling="2025-12-03 22:33:27.821247841 +0000 UTC m=+3104.563958436" observedRunningTime="2025-12-03 22:33:29.229066983 +0000 UTC m=+3105.971777578" watchObservedRunningTime="2025-12-03 22:33:29.234573154 +0000 UTC m=+3105.977283789" Dec 03 22:33:38 crc kubenswrapper[4715]: I1203 22:33:38.634780 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:33:38 crc kubenswrapper[4715]: E1203 22:33:38.635527 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:33:49 crc kubenswrapper[4715]: I1203 22:33:49.636889 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:33:49 crc kubenswrapper[4715]: E1203 22:33:49.637693 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:34:04 crc kubenswrapper[4715]: I1203 22:34:04.634825 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:34:04 crc kubenswrapper[4715]: E1203 22:34:04.635693 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:34:17 crc kubenswrapper[4715]: I1203 22:34:17.694079 4715 generic.go:334] "Generic (PLEG): container finished" podID="c3439842-caec-430b-a91e-0b03443eff15" containerID="1a4479a4e3020a484bd66101c82b779f4b8e606e932f2caa0ffe54fd099930db" exitCode=0 Dec 03 22:34:17 crc kubenswrapper[4715]: I1203 22:34:17.694201 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" event={"ID":"c3439842-caec-430b-a91e-0b03443eff15","Type":"ContainerDied","Data":"1a4479a4e3020a484bd66101c82b779f4b8e606e932f2caa0ffe54fd099930db"} Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.135984 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.204690 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-ovn-metadata-agent-neutron-config-0\") pod \"c3439842-caec-430b-a91e-0b03443eff15\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.204802 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-metadata-combined-ca-bundle\") pod \"c3439842-caec-430b-a91e-0b03443eff15\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.204856 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-ssh-key\") pod \"c3439842-caec-430b-a91e-0b03443eff15\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.204877 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-nova-metadata-neutron-config-0\") pod \"c3439842-caec-430b-a91e-0b03443eff15\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.204974 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-inventory\") pod \"c3439842-caec-430b-a91e-0b03443eff15\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.205036 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl7lz\" (UniqueName: \"kubernetes.io/projected/c3439842-caec-430b-a91e-0b03443eff15-kube-api-access-cl7lz\") pod \"c3439842-caec-430b-a91e-0b03443eff15\" (UID: \"c3439842-caec-430b-a91e-0b03443eff15\") " Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.210483 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c3439842-caec-430b-a91e-0b03443eff15" (UID: "c3439842-caec-430b-a91e-0b03443eff15"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.221142 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3439842-caec-430b-a91e-0b03443eff15-kube-api-access-cl7lz" (OuterVolumeSpecName: "kube-api-access-cl7lz") pod "c3439842-caec-430b-a91e-0b03443eff15" (UID: "c3439842-caec-430b-a91e-0b03443eff15"). InnerVolumeSpecName "kube-api-access-cl7lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.235280 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "c3439842-caec-430b-a91e-0b03443eff15" (UID: "c3439842-caec-430b-a91e-0b03443eff15"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.236464 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c3439842-caec-430b-a91e-0b03443eff15" (UID: "c3439842-caec-430b-a91e-0b03443eff15"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.245351 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "c3439842-caec-430b-a91e-0b03443eff15" (UID: "c3439842-caec-430b-a91e-0b03443eff15"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.250534 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-inventory" (OuterVolumeSpecName: "inventory") pod "c3439842-caec-430b-a91e-0b03443eff15" (UID: "c3439842-caec-430b-a91e-0b03443eff15"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.307102 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.307139 4715 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.307151 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.307162 4715 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.307171 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c3439842-caec-430b-a91e-0b03443eff15-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.307179 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl7lz\" (UniqueName: \"kubernetes.io/projected/c3439842-caec-430b-a91e-0b03443eff15-kube-api-access-cl7lz\") on node \"crc\" DevicePath \"\"" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.634562 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:34:19 crc kubenswrapper[4715]: E1203 22:34:19.635081 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.715473 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" event={"ID":"c3439842-caec-430b-a91e-0b03443eff15","Type":"ContainerDied","Data":"1e0b2157d11051af96d5a2995c7191e84d8777145ece8a54ad5e2c8165d50996"} Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.715566 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0b2157d11051af96d5a2995c7191e84d8777145ece8a54ad5e2c8165d50996" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.715652 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.804620 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4"] Dec 03 22:34:19 crc kubenswrapper[4715]: E1203 22:34:19.804996 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3439842-caec-430b-a91e-0b03443eff15" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.805012 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3439842-caec-430b-a91e-0b03443eff15" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.805192 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3439842-caec-430b-a91e-0b03443eff15" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.805807 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.808843 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.811092 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.811276 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.811407 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.811550 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.831126 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4"] Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.926049 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.926095 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.926182 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkzbw\" (UniqueName: \"kubernetes.io/projected/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-kube-api-access-fkzbw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.926388 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:19 crc kubenswrapper[4715]: I1203 22:34:19.926732 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.028690 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.028853 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.028961 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.028986 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.029014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkzbw\" (UniqueName: \"kubernetes.io/projected/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-kube-api-access-fkzbw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.033827 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.036409 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.039867 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.043683 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.056775 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkzbw\" (UniqueName: \"kubernetes.io/projected/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-kube-api-access-fkzbw\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.144026 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.658094 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4"] Dec 03 22:34:20 crc kubenswrapper[4715]: I1203 22:34:20.723686 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" event={"ID":"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a","Type":"ContainerStarted","Data":"06a168e790a9508ed9e3f4419aed2a5cc9b40fbffe0f32bc0571ad30930e45cc"} Dec 03 22:34:21 crc kubenswrapper[4715]: I1203 22:34:21.735344 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" event={"ID":"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a","Type":"ContainerStarted","Data":"f1fd8640fd3c1ab5ef1eb6c467ec55f1730e648ac68177bc4bd546e030bb6955"} Dec 03 22:34:21 crc kubenswrapper[4715]: I1203 22:34:21.754769 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" podStartSLOduration=2.337316823 podStartE2EDuration="2.754752923s" podCreationTimestamp="2025-12-03 22:34:19 +0000 UTC" firstStartedPulling="2025-12-03 22:34:20.662850771 +0000 UTC m=+3157.405561366" lastFinishedPulling="2025-12-03 22:34:21.080286861 +0000 UTC m=+3157.822997466" observedRunningTime="2025-12-03 22:34:21.750456285 +0000 UTC m=+3158.493166900" watchObservedRunningTime="2025-12-03 22:34:21.754752923 +0000 UTC m=+3158.497463518" Dec 03 22:34:32 crc kubenswrapper[4715]: I1203 22:34:32.634242 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:34:32 crc kubenswrapper[4715]: E1203 22:34:32.635028 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:34:44 crc kubenswrapper[4715]: I1203 22:34:44.635123 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:34:44 crc kubenswrapper[4715]: E1203 22:34:44.636317 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:34:58 crc kubenswrapper[4715]: I1203 22:34:58.634646 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:34:58 crc kubenswrapper[4715]: E1203 22:34:58.635260 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:35:12 crc kubenswrapper[4715]: I1203 22:35:12.633897 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:35:12 crc kubenswrapper[4715]: E1203 22:35:12.634772 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:35:25 crc kubenswrapper[4715]: I1203 22:35:25.634871 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:35:25 crc kubenswrapper[4715]: E1203 22:35:25.635824 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:35:40 crc kubenswrapper[4715]: I1203 22:35:40.657074 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:35:41 crc kubenswrapper[4715]: I1203 22:35:41.437195 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"7d9bfa2c1d02572cf252e696348da9eec449ec747bd736a0318c44de8c192b00"} Dec 03 22:38:05 crc kubenswrapper[4715]: I1203 22:38:05.160253 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:38:05 crc kubenswrapper[4715]: I1203 22:38:05.160800 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:38:35 crc kubenswrapper[4715]: I1203 22:38:35.160459 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:38:35 crc kubenswrapper[4715]: I1203 22:38:35.162833 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:38:39 crc kubenswrapper[4715]: I1203 22:38:39.147514 4715 generic.go:334] "Generic (PLEG): container finished" podID="e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" containerID="f1fd8640fd3c1ab5ef1eb6c467ec55f1730e648ac68177bc4bd546e030bb6955" exitCode=0 Dec 03 22:38:39 crc kubenswrapper[4715]: I1203 22:38:39.147589 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" event={"ID":"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a","Type":"ContainerDied","Data":"f1fd8640fd3c1ab5ef1eb6c467ec55f1730e648ac68177bc4bd546e030bb6955"} Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.640124 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.786456 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-secret-0\") pod \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.786616 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-inventory\") pod \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.786708 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-combined-ca-bundle\") pod \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.786784 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-ssh-key\") pod \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.786804 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkzbw\" (UniqueName: \"kubernetes.io/projected/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-kube-api-access-fkzbw\") pod \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\" (UID: \"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a\") " Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.792423 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" (UID: "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.794382 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-kube-api-access-fkzbw" (OuterVolumeSpecName: "kube-api-access-fkzbw") pod "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" (UID: "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a"). InnerVolumeSpecName "kube-api-access-fkzbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.815613 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" (UID: "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.816382 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" (UID: "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.840280 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-inventory" (OuterVolumeSpecName: "inventory") pod "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" (UID: "e419d6c2-a6b7-474d-aea8-6f90eb7ce85a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.889153 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.889181 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkzbw\" (UniqueName: \"kubernetes.io/projected/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-kube-api-access-fkzbw\") on node \"crc\" DevicePath \"\"" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.889194 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.889204 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:38:40 crc kubenswrapper[4715]: I1203 22:38:40.889215 4715 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e419d6c2-a6b7-474d-aea8-6f90eb7ce85a-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.188720 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" event={"ID":"e419d6c2-a6b7-474d-aea8-6f90eb7ce85a","Type":"ContainerDied","Data":"06a168e790a9508ed9e3f4419aed2a5cc9b40fbffe0f32bc0571ad30930e45cc"} Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.188777 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="06a168e790a9508ed9e3f4419aed2a5cc9b40fbffe0f32bc0571ad30930e45cc" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.188813 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.294668 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd"] Dec 03 22:38:41 crc kubenswrapper[4715]: E1203 22:38:41.295441 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.295467 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.295733 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e419d6c2-a6b7-474d-aea8-6f90eb7ce85a" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.296543 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.300340 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.300464 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.300484 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.300480 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.300791 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.302129 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.307289 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.307804 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd"] Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.399643 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.399727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.400127 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.400237 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.400322 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.400361 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.400841 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.400961 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsgcx\" (UniqueName: \"kubernetes.io/projected/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-kube-api-access-dsgcx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.401218 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503062 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503165 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503209 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503260 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503323 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsgcx\" (UniqueName: \"kubernetes.io/projected/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-kube-api-access-dsgcx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503458 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503561 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503640 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.503758 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.504301 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.507953 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.508879 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.509069 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.510156 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.510652 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.510936 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.512289 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.524404 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsgcx\" (UniqueName: \"kubernetes.io/projected/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-kube-api-access-dsgcx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-njnmd\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:41 crc kubenswrapper[4715]: I1203 22:38:41.659422 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:38:42 crc kubenswrapper[4715]: I1203 22:38:42.242808 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd"] Dec 03 22:38:42 crc kubenswrapper[4715]: I1203 22:38:42.243677 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:38:43 crc kubenswrapper[4715]: I1203 22:38:43.223936 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" event={"ID":"10aa3b67-88a4-4a5d-9d81-096c5b944fb7","Type":"ContainerStarted","Data":"5e52f22e4ed961887ee4e0847ae4d39a9bb0c272256a22ac144a16ffb92cbdbc"} Dec 03 22:38:43 crc kubenswrapper[4715]: I1203 22:38:43.224231 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" event={"ID":"10aa3b67-88a4-4a5d-9d81-096c5b944fb7","Type":"ContainerStarted","Data":"b4e387c2cb4c4559e891db9b90b741782c7c86e2fd7b613450976204f6fcdd36"} Dec 03 22:38:43 crc kubenswrapper[4715]: I1203 22:38:43.242067 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" podStartSLOduration=1.759312194 podStartE2EDuration="2.242049271s" podCreationTimestamp="2025-12-03 22:38:41 +0000 UTC" firstStartedPulling="2025-12-03 22:38:42.241136529 +0000 UTC m=+3418.983847164" lastFinishedPulling="2025-12-03 22:38:42.723873626 +0000 UTC m=+3419.466584241" observedRunningTime="2025-12-03 22:38:43.241072624 +0000 UTC m=+3419.983783219" watchObservedRunningTime="2025-12-03 22:38:43.242049271 +0000 UTC m=+3419.984759866" Dec 03 22:39:05 crc kubenswrapper[4715]: I1203 22:39:05.160243 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:39:05 crc kubenswrapper[4715]: I1203 22:39:05.160879 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:39:05 crc kubenswrapper[4715]: I1203 22:39:05.160934 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:39:05 crc kubenswrapper[4715]: I1203 22:39:05.161922 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7d9bfa2c1d02572cf252e696348da9eec449ec747bd736a0318c44de8c192b00"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:39:05 crc kubenswrapper[4715]: I1203 22:39:05.161979 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://7d9bfa2c1d02572cf252e696348da9eec449ec747bd736a0318c44de8c192b00" gracePeriod=600 Dec 03 22:39:05 crc kubenswrapper[4715]: I1203 22:39:05.494238 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="7d9bfa2c1d02572cf252e696348da9eec449ec747bd736a0318c44de8c192b00" exitCode=0 Dec 03 22:39:05 crc kubenswrapper[4715]: I1203 22:39:05.494312 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"7d9bfa2c1d02572cf252e696348da9eec449ec747bd736a0318c44de8c192b00"} Dec 03 22:39:05 crc kubenswrapper[4715]: I1203 22:39:05.494786 4715 scope.go:117] "RemoveContainer" containerID="4ddccb9ec83b2acadd1a1e9b69dba69e15d9c3c5e76290a1228a2fea4695990a" Dec 03 22:39:06 crc kubenswrapper[4715]: I1203 22:39:06.507154 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe"} Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.726485 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-szxzb"] Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.729413 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.854838 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-catalog-content\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.855224 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhbj4\" (UniqueName: \"kubernetes.io/projected/36d4397d-5592-40b4-bdc3-020441d0c357-kube-api-access-fhbj4\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.855252 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-utilities\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.932056 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcdq"] Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.935487 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.957859 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-catalog-content\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.958052 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhbj4\" (UniqueName: \"kubernetes.io/projected/36d4397d-5592-40b4-bdc3-020441d0c357-kube-api-access-fhbj4\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.958156 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-utilities\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.958825 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-utilities\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.958969 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-catalog-content\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:36 crc kubenswrapper[4715]: I1203 22:39:36.993306 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhbj4\" (UniqueName: \"kubernetes.io/projected/36d4397d-5592-40b4-bdc3-020441d0c357-kube-api-access-fhbj4\") pod \"community-operators-szxzb\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.049198 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.060397 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-catalog-content\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.060655 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzxnn\" (UniqueName: \"kubernetes.io/projected/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-kube-api-access-tzxnn\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.060777 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-utilities\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.162489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-catalog-content\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.162625 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzxnn\" (UniqueName: \"kubernetes.io/projected/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-kube-api-access-tzxnn\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.162674 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-utilities\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.163029 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-catalog-content\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.163122 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-utilities\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.178613 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzxnn\" (UniqueName: \"kubernetes.io/projected/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-kube-api-access-tzxnn\") pod \"redhat-marketplace-gxcdq\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.254357 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.355574 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcdq"] Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.401125 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-szxzb"] Dec 03 22:39:37 crc kubenswrapper[4715]: I1203 22:39:37.942366 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-szxzb"] Dec 03 22:39:38 crc kubenswrapper[4715]: I1203 22:39:38.074176 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcdq"] Dec 03 22:39:38 crc kubenswrapper[4715]: I1203 22:39:38.347232 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcdq" event={"ID":"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02","Type":"ContainerStarted","Data":"900a5bd9409bfc86e44b0482e4d49549ea134fb6b177236d00e01a8435b40967"} Dec 03 22:39:38 crc kubenswrapper[4715]: I1203 22:39:38.348239 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szxzb" event={"ID":"36d4397d-5592-40b4-bdc3-020441d0c357","Type":"ContainerStarted","Data":"309de2c071e8f7aa11be669677cdaad01050c8783f2ca5634b4865aabef1092f"} Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.122840 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mfdjb"] Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.127030 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.148761 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mfdjb"] Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.217911 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-utilities\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.218057 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scg5q\" (UniqueName: \"kubernetes.io/projected/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-kube-api-access-scg5q\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.218248 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-catalog-content\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.320828 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scg5q\" (UniqueName: \"kubernetes.io/projected/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-kube-api-access-scg5q\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.321063 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-catalog-content\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.321231 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-utilities\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.322049 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-catalog-content\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.322182 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-utilities\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.346957 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scg5q\" (UniqueName: \"kubernetes.io/projected/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-kube-api-access-scg5q\") pod \"certified-operators-mfdjb\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.364427 4715 generic.go:334] "Generic (PLEG): container finished" podID="36d4397d-5592-40b4-bdc3-020441d0c357" containerID="ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16" exitCode=0 Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.366469 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szxzb" event={"ID":"36d4397d-5592-40b4-bdc3-020441d0c357","Type":"ContainerDied","Data":"ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16"} Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.374153 4715 generic.go:334] "Generic (PLEG): container finished" podID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerID="87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981" exitCode=0 Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.374229 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcdq" event={"ID":"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02","Type":"ContainerDied","Data":"87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981"} Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.453091 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:39 crc kubenswrapper[4715]: I1203 22:39:39.986278 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mfdjb"] Dec 03 22:39:40 crc kubenswrapper[4715]: I1203 22:39:40.386041 4715 generic.go:334] "Generic (PLEG): container finished" podID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerID="e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93" exitCode=0 Dec 03 22:39:40 crc kubenswrapper[4715]: I1203 22:39:40.386366 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcdq" event={"ID":"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02","Type":"ContainerDied","Data":"e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93"} Dec 03 22:39:40 crc kubenswrapper[4715]: I1203 22:39:40.389988 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szxzb" event={"ID":"36d4397d-5592-40b4-bdc3-020441d0c357","Type":"ContainerStarted","Data":"bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4"} Dec 03 22:39:40 crc kubenswrapper[4715]: I1203 22:39:40.392308 4715 generic.go:334] "Generic (PLEG): container finished" podID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerID="5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f" exitCode=0 Dec 03 22:39:40 crc kubenswrapper[4715]: I1203 22:39:40.392364 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfdjb" event={"ID":"ca94e79e-c93c-49d9-8b00-f21bbbd2139a","Type":"ContainerDied","Data":"5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f"} Dec 03 22:39:40 crc kubenswrapper[4715]: I1203 22:39:40.392394 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfdjb" event={"ID":"ca94e79e-c93c-49d9-8b00-f21bbbd2139a","Type":"ContainerStarted","Data":"a46b04bd8955651b438349400b66b7cdc8c56b0267cf86b4975d3fc838a7bf1e"} Dec 03 22:39:41 crc kubenswrapper[4715]: I1203 22:39:41.436360 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfdjb" event={"ID":"ca94e79e-c93c-49d9-8b00-f21bbbd2139a","Type":"ContainerStarted","Data":"04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac"} Dec 03 22:39:41 crc kubenswrapper[4715]: I1203 22:39:41.440803 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcdq" event={"ID":"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02","Type":"ContainerStarted","Data":"17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824"} Dec 03 22:39:41 crc kubenswrapper[4715]: I1203 22:39:41.444490 4715 generic.go:334] "Generic (PLEG): container finished" podID="36d4397d-5592-40b4-bdc3-020441d0c357" containerID="bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4" exitCode=0 Dec 03 22:39:41 crc kubenswrapper[4715]: I1203 22:39:41.444555 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szxzb" event={"ID":"36d4397d-5592-40b4-bdc3-020441d0c357","Type":"ContainerDied","Data":"bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4"} Dec 03 22:39:41 crc kubenswrapper[4715]: I1203 22:39:41.484670 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gxcdq" podStartSLOduration=4.07812529 podStartE2EDuration="5.484645382s" podCreationTimestamp="2025-12-03 22:39:36 +0000 UTC" firstStartedPulling="2025-12-03 22:39:39.378575559 +0000 UTC m=+3476.121286154" lastFinishedPulling="2025-12-03 22:39:40.785095611 +0000 UTC m=+3477.527806246" observedRunningTime="2025-12-03 22:39:41.48310662 +0000 UTC m=+3478.225817225" watchObservedRunningTime="2025-12-03 22:39:41.484645382 +0000 UTC m=+3478.227355997" Dec 03 22:39:43 crc kubenswrapper[4715]: I1203 22:39:43.466645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szxzb" event={"ID":"36d4397d-5592-40b4-bdc3-020441d0c357","Type":"ContainerStarted","Data":"c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5"} Dec 03 22:39:43 crc kubenswrapper[4715]: I1203 22:39:43.472027 4715 generic.go:334] "Generic (PLEG): container finished" podID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerID="04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac" exitCode=0 Dec 03 22:39:43 crc kubenswrapper[4715]: I1203 22:39:43.472069 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfdjb" event={"ID":"ca94e79e-c93c-49d9-8b00-f21bbbd2139a","Type":"ContainerDied","Data":"04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac"} Dec 03 22:39:43 crc kubenswrapper[4715]: I1203 22:39:43.489012 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-szxzb" podStartSLOduration=4.489402934 podStartE2EDuration="7.488988283s" podCreationTimestamp="2025-12-03 22:39:36 +0000 UTC" firstStartedPulling="2025-12-03 22:39:39.369694909 +0000 UTC m=+3476.112405504" lastFinishedPulling="2025-12-03 22:39:42.369280238 +0000 UTC m=+3479.111990853" observedRunningTime="2025-12-03 22:39:43.482804156 +0000 UTC m=+3480.225514751" watchObservedRunningTime="2025-12-03 22:39:43.488988283 +0000 UTC m=+3480.231698878" Dec 03 22:39:45 crc kubenswrapper[4715]: I1203 22:39:45.513487 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfdjb" event={"ID":"ca94e79e-c93c-49d9-8b00-f21bbbd2139a","Type":"ContainerStarted","Data":"a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902"} Dec 03 22:39:45 crc kubenswrapper[4715]: I1203 22:39:45.529749 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mfdjb" podStartSLOduration=1.650928714 podStartE2EDuration="6.529733929s" podCreationTimestamp="2025-12-03 22:39:39 +0000 UTC" firstStartedPulling="2025-12-03 22:39:40.394271121 +0000 UTC m=+3477.136981716" lastFinishedPulling="2025-12-03 22:39:45.273076166 +0000 UTC m=+3482.015786931" observedRunningTime="2025-12-03 22:39:45.528927026 +0000 UTC m=+3482.271637611" watchObservedRunningTime="2025-12-03 22:39:45.529733929 +0000 UTC m=+3482.272444524" Dec 03 22:39:47 crc kubenswrapper[4715]: I1203 22:39:47.049444 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:47 crc kubenswrapper[4715]: I1203 22:39:47.050695 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:47 crc kubenswrapper[4715]: I1203 22:39:47.097216 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:47 crc kubenswrapper[4715]: I1203 22:39:47.255426 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:47 crc kubenswrapper[4715]: I1203 22:39:47.255496 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:47 crc kubenswrapper[4715]: I1203 22:39:47.316656 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:47 crc kubenswrapper[4715]: I1203 22:39:47.584982 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:47 crc kubenswrapper[4715]: I1203 22:39:47.585811 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:48 crc kubenswrapper[4715]: I1203 22:39:48.718385 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-szxzb"] Dec 03 22:39:49 crc kubenswrapper[4715]: I1203 22:39:49.454958 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:49 crc kubenswrapper[4715]: I1203 22:39:49.455038 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:49 crc kubenswrapper[4715]: I1203 22:39:49.530150 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:49 crc kubenswrapper[4715]: I1203 22:39:49.545215 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-szxzb" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" containerName="registry-server" containerID="cri-o://c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5" gracePeriod=2 Dec 03 22:39:49 crc kubenswrapper[4715]: I1203 22:39:49.717272 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcdq"] Dec 03 22:39:49 crc kubenswrapper[4715]: I1203 22:39:49.717706 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gxcdq" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerName="registry-server" containerID="cri-o://17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824" gracePeriod=2 Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.050009 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.144332 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-utilities\") pod \"36d4397d-5592-40b4-bdc3-020441d0c357\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.144452 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhbj4\" (UniqueName: \"kubernetes.io/projected/36d4397d-5592-40b4-bdc3-020441d0c357-kube-api-access-fhbj4\") pod \"36d4397d-5592-40b4-bdc3-020441d0c357\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.144537 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-catalog-content\") pod \"36d4397d-5592-40b4-bdc3-020441d0c357\" (UID: \"36d4397d-5592-40b4-bdc3-020441d0c357\") " Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.145412 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-utilities" (OuterVolumeSpecName: "utilities") pod "36d4397d-5592-40b4-bdc3-020441d0c357" (UID: "36d4397d-5592-40b4-bdc3-020441d0c357"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.151222 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d4397d-5592-40b4-bdc3-020441d0c357-kube-api-access-fhbj4" (OuterVolumeSpecName: "kube-api-access-fhbj4") pod "36d4397d-5592-40b4-bdc3-020441d0c357" (UID: "36d4397d-5592-40b4-bdc3-020441d0c357"). InnerVolumeSpecName "kube-api-access-fhbj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.156328 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.198977 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "36d4397d-5592-40b4-bdc3-020441d0c357" (UID: "36d4397d-5592-40b4-bdc3-020441d0c357"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.246345 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzxnn\" (UniqueName: \"kubernetes.io/projected/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-kube-api-access-tzxnn\") pod \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.246578 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-catalog-content\") pod \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.246689 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-utilities\") pod \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\" (UID: \"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02\") " Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.247336 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhbj4\" (UniqueName: \"kubernetes.io/projected/36d4397d-5592-40b4-bdc3-020441d0c357-kube-api-access-fhbj4\") on node \"crc\" DevicePath \"\"" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.247361 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.247378 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36d4397d-5592-40b4-bdc3-020441d0c357-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.247348 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-utilities" (OuterVolumeSpecName: "utilities") pod "2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" (UID: "2e8a2b5b-7385-404b-b71d-6d84fe9a1c02"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.249629 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-kube-api-access-tzxnn" (OuterVolumeSpecName: "kube-api-access-tzxnn") pod "2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" (UID: "2e8a2b5b-7385-404b-b71d-6d84fe9a1c02"). InnerVolumeSpecName "kube-api-access-tzxnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.264274 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" (UID: "2e8a2b5b-7385-404b-b71d-6d84fe9a1c02"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.349319 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.349378 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzxnn\" (UniqueName: \"kubernetes.io/projected/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-kube-api-access-tzxnn\") on node \"crc\" DevicePath \"\"" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.349405 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.557220 4715 generic.go:334] "Generic (PLEG): container finished" podID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerID="17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824" exitCode=0 Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.557296 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcdq" event={"ID":"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02","Type":"ContainerDied","Data":"17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824"} Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.557351 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gxcdq" event={"ID":"2e8a2b5b-7385-404b-b71d-6d84fe9a1c02","Type":"ContainerDied","Data":"900a5bd9409bfc86e44b0482e4d49549ea134fb6b177236d00e01a8435b40967"} Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.557370 4715 scope.go:117] "RemoveContainer" containerID="17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.557304 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gxcdq" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.561355 4715 generic.go:334] "Generic (PLEG): container finished" podID="36d4397d-5592-40b4-bdc3-020441d0c357" containerID="c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5" exitCode=0 Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.561394 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szxzb" event={"ID":"36d4397d-5592-40b4-bdc3-020441d0c357","Type":"ContainerDied","Data":"c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5"} Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.561425 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szxzb" event={"ID":"36d4397d-5592-40b4-bdc3-020441d0c357","Type":"ContainerDied","Data":"309de2c071e8f7aa11be669677cdaad01050c8783f2ca5634b4865aabef1092f"} Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.561483 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szxzb" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.581178 4715 scope.go:117] "RemoveContainer" containerID="e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.597993 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcdq"] Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.616747 4715 scope.go:117] "RemoveContainer" containerID="87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.616934 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gxcdq"] Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.625957 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-szxzb"] Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.634977 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-szxzb"] Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.639489 4715 scope.go:117] "RemoveContainer" containerID="17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824" Dec 03 22:39:50 crc kubenswrapper[4715]: E1203 22:39:50.640036 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824\": container with ID starting with 17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824 not found: ID does not exist" containerID="17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.640239 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824"} err="failed to get container status \"17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824\": rpc error: code = NotFound desc = could not find container \"17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824\": container with ID starting with 17fd92eb398cfb7a45bbc463ad0da74ff471f591cf8896274683d9321de6f824 not found: ID does not exist" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.640397 4715 scope.go:117] "RemoveContainer" containerID="e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93" Dec 03 22:39:50 crc kubenswrapper[4715]: E1203 22:39:50.649325 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93\": container with ID starting with e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93 not found: ID does not exist" containerID="e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.649374 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93"} err="failed to get container status \"e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93\": rpc error: code = NotFound desc = could not find container \"e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93\": container with ID starting with e42ea077181d5a3b744e3913ec8e6749f09dbd1e19fe06d0995c433c44a92b93 not found: ID does not exist" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.649397 4715 scope.go:117] "RemoveContainer" containerID="87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981" Dec 03 22:39:50 crc kubenswrapper[4715]: E1203 22:39:50.649906 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981\": container with ID starting with 87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981 not found: ID does not exist" containerID="87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.649958 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981"} err="failed to get container status \"87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981\": rpc error: code = NotFound desc = could not find container \"87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981\": container with ID starting with 87ac8ae05aaab712e3a7f2db3b9300d6e162db9452dea07b0673153017b25981 not found: ID does not exist" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.649989 4715 scope.go:117] "RemoveContainer" containerID="c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.671322 4715 scope.go:117] "RemoveContainer" containerID="bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.732374 4715 scope.go:117] "RemoveContainer" containerID="ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.781893 4715 scope.go:117] "RemoveContainer" containerID="c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5" Dec 03 22:39:50 crc kubenswrapper[4715]: E1203 22:39:50.782697 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5\": container with ID starting with c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5 not found: ID does not exist" containerID="c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.782753 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5"} err="failed to get container status \"c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5\": rpc error: code = NotFound desc = could not find container \"c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5\": container with ID starting with c2f94a72238b3066bae3058071e35d925ffc878cb5b78c80c7940fbe6bae05d5 not found: ID does not exist" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.782791 4715 scope.go:117] "RemoveContainer" containerID="bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4" Dec 03 22:39:50 crc kubenswrapper[4715]: E1203 22:39:50.783180 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4\": container with ID starting with bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4 not found: ID does not exist" containerID="bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.783353 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4"} err="failed to get container status \"bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4\": rpc error: code = NotFound desc = could not find container \"bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4\": container with ID starting with bac8103bdc71d831c483bda66305a95430af594e2ced2cd8deeef542e8959ab4 not found: ID does not exist" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.783535 4715 scope.go:117] "RemoveContainer" containerID="ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16" Dec 03 22:39:50 crc kubenswrapper[4715]: E1203 22:39:50.784135 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16\": container with ID starting with ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16 not found: ID does not exist" containerID="ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16" Dec 03 22:39:50 crc kubenswrapper[4715]: I1203 22:39:50.784168 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16"} err="failed to get container status \"ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16\": rpc error: code = NotFound desc = could not find container \"ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16\": container with ID starting with ee7b940610b7f6f441b41dfd47d72d12e76e6c954400608b1e13e52a80045f16 not found: ID does not exist" Dec 03 22:39:51 crc kubenswrapper[4715]: I1203 22:39:51.651157 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" path="/var/lib/kubelet/pods/2e8a2b5b-7385-404b-b71d-6d84fe9a1c02/volumes" Dec 03 22:39:51 crc kubenswrapper[4715]: I1203 22:39:51.652518 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" path="/var/lib/kubelet/pods/36d4397d-5592-40b4-bdc3-020441d0c357/volumes" Dec 03 22:39:59 crc kubenswrapper[4715]: I1203 22:39:59.508777 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:39:59 crc kubenswrapper[4715]: I1203 22:39:59.564208 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mfdjb"] Dec 03 22:39:59 crc kubenswrapper[4715]: I1203 22:39:59.657482 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mfdjb" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerName="registry-server" containerID="cri-o://a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902" gracePeriod=2 Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.114854 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.231935 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scg5q\" (UniqueName: \"kubernetes.io/projected/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-kube-api-access-scg5q\") pod \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.232034 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-catalog-content\") pod \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.232239 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-utilities\") pod \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\" (UID: \"ca94e79e-c93c-49d9-8b00-f21bbbd2139a\") " Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.233385 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-utilities" (OuterVolumeSpecName: "utilities") pod "ca94e79e-c93c-49d9-8b00-f21bbbd2139a" (UID: "ca94e79e-c93c-49d9-8b00-f21bbbd2139a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.237358 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-kube-api-access-scg5q" (OuterVolumeSpecName: "kube-api-access-scg5q") pod "ca94e79e-c93c-49d9-8b00-f21bbbd2139a" (UID: "ca94e79e-c93c-49d9-8b00-f21bbbd2139a"). InnerVolumeSpecName "kube-api-access-scg5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.278812 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca94e79e-c93c-49d9-8b00-f21bbbd2139a" (UID: "ca94e79e-c93c-49d9-8b00-f21bbbd2139a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.334861 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scg5q\" (UniqueName: \"kubernetes.io/projected/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-kube-api-access-scg5q\") on node \"crc\" DevicePath \"\"" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.334896 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.334908 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca94e79e-c93c-49d9-8b00-f21bbbd2139a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.668296 4715 generic.go:334] "Generic (PLEG): container finished" podID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerID="a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902" exitCode=0 Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.668342 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfdjb" event={"ID":"ca94e79e-c93c-49d9-8b00-f21bbbd2139a","Type":"ContainerDied","Data":"a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902"} Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.668378 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mfdjb" event={"ID":"ca94e79e-c93c-49d9-8b00-f21bbbd2139a","Type":"ContainerDied","Data":"a46b04bd8955651b438349400b66b7cdc8c56b0267cf86b4975d3fc838a7bf1e"} Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.668396 4715 scope.go:117] "RemoveContainer" containerID="a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.668409 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mfdjb" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.709858 4715 scope.go:117] "RemoveContainer" containerID="04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.716167 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mfdjb"] Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.726099 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mfdjb"] Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.737069 4715 scope.go:117] "RemoveContainer" containerID="5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.787303 4715 scope.go:117] "RemoveContainer" containerID="a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902" Dec 03 22:40:00 crc kubenswrapper[4715]: E1203 22:40:00.787812 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902\": container with ID starting with a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902 not found: ID does not exist" containerID="a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.787864 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902"} err="failed to get container status \"a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902\": rpc error: code = NotFound desc = could not find container \"a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902\": container with ID starting with a6e0c748bdba2c71cbfbc54ea31ae4d68e14e9924706926128a18fb943e93902 not found: ID does not exist" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.787898 4715 scope.go:117] "RemoveContainer" containerID="04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac" Dec 03 22:40:00 crc kubenswrapper[4715]: E1203 22:40:00.788237 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac\": container with ID starting with 04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac not found: ID does not exist" containerID="04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.788271 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac"} err="failed to get container status \"04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac\": rpc error: code = NotFound desc = could not find container \"04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac\": container with ID starting with 04a2391abab03ac2fc5f8e11ef2510e733d3223ad1f6e233ee4a9f5090a225ac not found: ID does not exist" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.788297 4715 scope.go:117] "RemoveContainer" containerID="5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f" Dec 03 22:40:00 crc kubenswrapper[4715]: E1203 22:40:00.788557 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f\": container with ID starting with 5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f not found: ID does not exist" containerID="5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f" Dec 03 22:40:00 crc kubenswrapper[4715]: I1203 22:40:00.788681 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f"} err="failed to get container status \"5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f\": rpc error: code = NotFound desc = could not find container \"5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f\": container with ID starting with 5020d0f6cc69a5614dbcf432f1cf3acdda541ea44af96d88b3e2b0cecbdacb7f not found: ID does not exist" Dec 03 22:40:01 crc kubenswrapper[4715]: I1203 22:40:01.649854 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" path="/var/lib/kubelet/pods/ca94e79e-c93c-49d9-8b00-f21bbbd2139a/volumes" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.342329 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9879p"] Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343194 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerName="extract-content" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343206 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerName="extract-content" Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343218 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" containerName="extract-utilities" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343224 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" containerName="extract-utilities" Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343246 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerName="extract-utilities" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343255 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerName="extract-utilities" Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343269 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343274 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343286 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerName="extract-content" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343292 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerName="extract-content" Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343302 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343308 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343317 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343322 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343331 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerName="extract-utilities" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343337 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerName="extract-utilities" Dec 03 22:40:39 crc kubenswrapper[4715]: E1203 22:40:39.343349 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" containerName="extract-content" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343354 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" containerName="extract-content" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343567 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d4397d-5592-40b4-bdc3-020441d0c357" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343585 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e8a2b5b-7385-404b-b71d-6d84fe9a1c02" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.343607 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca94e79e-c93c-49d9-8b00-f21bbbd2139a" containerName="registry-server" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.345111 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.374388 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9879p"] Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.436408 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxsvh\" (UniqueName: \"kubernetes.io/projected/16bdae06-b96e-49f9-afce-21ccf312277d-kube-api-access-bxsvh\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.436727 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-catalog-content\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.436959 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-utilities\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.538405 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-utilities\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.538488 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxsvh\" (UniqueName: \"kubernetes.io/projected/16bdae06-b96e-49f9-afce-21ccf312277d-kube-api-access-bxsvh\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.538557 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-catalog-content\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.539661 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-utilities\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.540127 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-catalog-content\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.560460 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxsvh\" (UniqueName: \"kubernetes.io/projected/16bdae06-b96e-49f9-afce-21ccf312277d-kube-api-access-bxsvh\") pod \"redhat-operators-9879p\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:39 crc kubenswrapper[4715]: I1203 22:40:39.666619 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:40 crc kubenswrapper[4715]: I1203 22:40:40.117522 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9879p"] Dec 03 22:40:40 crc kubenswrapper[4715]: W1203 22:40:40.122436 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16bdae06_b96e_49f9_afce_21ccf312277d.slice/crio-52dabbcfe794db52c14185d01f87a639f7f058c1a824f42a1050b874772504e8 WatchSource:0}: Error finding container 52dabbcfe794db52c14185d01f87a639f7f058c1a824f42a1050b874772504e8: Status 404 returned error can't find the container with id 52dabbcfe794db52c14185d01f87a639f7f058c1a824f42a1050b874772504e8 Dec 03 22:40:41 crc kubenswrapper[4715]: I1203 22:40:41.044887 4715 generic.go:334] "Generic (PLEG): container finished" podID="16bdae06-b96e-49f9-afce-21ccf312277d" containerID="b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e" exitCode=0 Dec 03 22:40:41 crc kubenswrapper[4715]: I1203 22:40:41.044993 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9879p" event={"ID":"16bdae06-b96e-49f9-afce-21ccf312277d","Type":"ContainerDied","Data":"b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e"} Dec 03 22:40:41 crc kubenswrapper[4715]: I1203 22:40:41.045182 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9879p" event={"ID":"16bdae06-b96e-49f9-afce-21ccf312277d","Type":"ContainerStarted","Data":"52dabbcfe794db52c14185d01f87a639f7f058c1a824f42a1050b874772504e8"} Dec 03 22:40:43 crc kubenswrapper[4715]: I1203 22:40:43.265852 4715 generic.go:334] "Generic (PLEG): container finished" podID="16bdae06-b96e-49f9-afce-21ccf312277d" containerID="80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b" exitCode=0 Dec 03 22:40:43 crc kubenswrapper[4715]: I1203 22:40:43.266224 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9879p" event={"ID":"16bdae06-b96e-49f9-afce-21ccf312277d","Type":"ContainerDied","Data":"80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b"} Dec 03 22:40:44 crc kubenswrapper[4715]: I1203 22:40:44.277134 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9879p" event={"ID":"16bdae06-b96e-49f9-afce-21ccf312277d","Type":"ContainerStarted","Data":"a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98"} Dec 03 22:40:44 crc kubenswrapper[4715]: I1203 22:40:44.302747 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9879p" podStartSLOduration=2.653078762 podStartE2EDuration="5.302724176s" podCreationTimestamp="2025-12-03 22:40:39 +0000 UTC" firstStartedPulling="2025-12-03 22:40:41.048287364 +0000 UTC m=+3537.790997979" lastFinishedPulling="2025-12-03 22:40:43.697932798 +0000 UTC m=+3540.440643393" observedRunningTime="2025-12-03 22:40:44.296458667 +0000 UTC m=+3541.039169302" watchObservedRunningTime="2025-12-03 22:40:44.302724176 +0000 UTC m=+3541.045434771" Dec 03 22:40:49 crc kubenswrapper[4715]: I1203 22:40:49.667734 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:49 crc kubenswrapper[4715]: I1203 22:40:49.668294 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:49 crc kubenswrapper[4715]: I1203 22:40:49.730627 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:50 crc kubenswrapper[4715]: I1203 22:40:50.401264 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:50 crc kubenswrapper[4715]: I1203 22:40:50.446013 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9879p"] Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.347168 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9879p" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" containerName="registry-server" containerID="cri-o://a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98" gracePeriod=2 Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.856540 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.861582 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxsvh\" (UniqueName: \"kubernetes.io/projected/16bdae06-b96e-49f9-afce-21ccf312277d-kube-api-access-bxsvh\") pod \"16bdae06-b96e-49f9-afce-21ccf312277d\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.861705 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-utilities\") pod \"16bdae06-b96e-49f9-afce-21ccf312277d\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.861741 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-catalog-content\") pod \"16bdae06-b96e-49f9-afce-21ccf312277d\" (UID: \"16bdae06-b96e-49f9-afce-21ccf312277d\") " Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.862965 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-utilities" (OuterVolumeSpecName: "utilities") pod "16bdae06-b96e-49f9-afce-21ccf312277d" (UID: "16bdae06-b96e-49f9-afce-21ccf312277d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.869715 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16bdae06-b96e-49f9-afce-21ccf312277d-kube-api-access-bxsvh" (OuterVolumeSpecName: "kube-api-access-bxsvh") pod "16bdae06-b96e-49f9-afce-21ccf312277d" (UID: "16bdae06-b96e-49f9-afce-21ccf312277d"). InnerVolumeSpecName "kube-api-access-bxsvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.964681 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxsvh\" (UniqueName: \"kubernetes.io/projected/16bdae06-b96e-49f9-afce-21ccf312277d-kube-api-access-bxsvh\") on node \"crc\" DevicePath \"\"" Dec 03 22:40:52 crc kubenswrapper[4715]: I1203 22:40:52.964728 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.005130 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "16bdae06-b96e-49f9-afce-21ccf312277d" (UID: "16bdae06-b96e-49f9-afce-21ccf312277d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.065699 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16bdae06-b96e-49f9-afce-21ccf312277d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.360262 4715 generic.go:334] "Generic (PLEG): container finished" podID="16bdae06-b96e-49f9-afce-21ccf312277d" containerID="a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98" exitCode=0 Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.360319 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9879p" event={"ID":"16bdae06-b96e-49f9-afce-21ccf312277d","Type":"ContainerDied","Data":"a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98"} Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.360359 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9879p" event={"ID":"16bdae06-b96e-49f9-afce-21ccf312277d","Type":"ContainerDied","Data":"52dabbcfe794db52c14185d01f87a639f7f058c1a824f42a1050b874772504e8"} Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.360365 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9879p" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.360381 4715 scope.go:117] "RemoveContainer" containerID="a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.391423 4715 scope.go:117] "RemoveContainer" containerID="80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.392475 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9879p"] Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.401576 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9879p"] Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.417494 4715 scope.go:117] "RemoveContainer" containerID="b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.474961 4715 scope.go:117] "RemoveContainer" containerID="a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98" Dec 03 22:40:53 crc kubenswrapper[4715]: E1203 22:40:53.475376 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98\": container with ID starting with a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98 not found: ID does not exist" containerID="a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.475426 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98"} err="failed to get container status \"a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98\": rpc error: code = NotFound desc = could not find container \"a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98\": container with ID starting with a26e98cbc9b466f07b12f7c4673e5ea102ea152a1e065bc5f5cc6e247d1b9c98 not found: ID does not exist" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.475455 4715 scope.go:117] "RemoveContainer" containerID="80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b" Dec 03 22:40:53 crc kubenswrapper[4715]: E1203 22:40:53.475848 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b\": container with ID starting with 80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b not found: ID does not exist" containerID="80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.475876 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b"} err="failed to get container status \"80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b\": rpc error: code = NotFound desc = could not find container \"80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b\": container with ID starting with 80dc103328a2ebc5ff5c3a8b079f66b1b806393765c9b93fc973619615f4b12b not found: ID does not exist" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.475895 4715 scope.go:117] "RemoveContainer" containerID="b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e" Dec 03 22:40:53 crc kubenswrapper[4715]: E1203 22:40:53.476242 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e\": container with ID starting with b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e not found: ID does not exist" containerID="b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.476271 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e"} err="failed to get container status \"b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e\": rpc error: code = NotFound desc = could not find container \"b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e\": container with ID starting with b33f07b08a713aed38e7982afae8afb4c4bb1e0edb9b5c966b7191b8bf19a88e not found: ID does not exist" Dec 03 22:40:53 crc kubenswrapper[4715]: I1203 22:40:53.648638 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" path="/var/lib/kubelet/pods/16bdae06-b96e-49f9-afce-21ccf312277d/volumes" Dec 03 22:41:05 crc kubenswrapper[4715]: I1203 22:41:05.161322 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:41:05 crc kubenswrapper[4715]: I1203 22:41:05.162402 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:41:28 crc kubenswrapper[4715]: I1203 22:41:28.716845 4715 generic.go:334] "Generic (PLEG): container finished" podID="10aa3b67-88a4-4a5d-9d81-096c5b944fb7" containerID="5e52f22e4ed961887ee4e0847ae4d39a9bb0c272256a22ac144a16ffb92cbdbc" exitCode=0 Dec 03 22:41:28 crc kubenswrapper[4715]: I1203 22:41:28.717710 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" event={"ID":"10aa3b67-88a4-4a5d-9d81-096c5b944fb7","Type":"ContainerDied","Data":"5e52f22e4ed961887ee4e0847ae4d39a9bb0c272256a22ac144a16ffb92cbdbc"} Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.201414 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.333977 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-extra-config-0\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.334055 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-ssh-key\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.334243 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsgcx\" (UniqueName: \"kubernetes.io/projected/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-kube-api-access-dsgcx\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.334288 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-0\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.334327 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-inventory\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.334372 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-1\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.334401 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-1\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.334441 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-0\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.334472 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-combined-ca-bundle\") pod \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\" (UID: \"10aa3b67-88a4-4a5d-9d81-096c5b944fb7\") " Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.341748 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.355904 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-kube-api-access-dsgcx" (OuterVolumeSpecName: "kube-api-access-dsgcx") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "kube-api-access-dsgcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.365369 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.379887 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.381185 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.382410 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-inventory" (OuterVolumeSpecName: "inventory") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.384114 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.390823 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.390854 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "10aa3b67-88a4-4a5d-9d81-096c5b944fb7" (UID: "10aa3b67-88a4-4a5d-9d81-096c5b944fb7"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.437347 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsgcx\" (UniqueName: \"kubernetes.io/projected/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-kube-api-access-dsgcx\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.437652 4715 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.437779 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.437864 4715 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.437939 4715 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.438029 4715 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.438116 4715 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.438192 4715 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.438278 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/10aa3b67-88a4-4a5d-9d81-096c5b944fb7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.739667 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" event={"ID":"10aa3b67-88a4-4a5d-9d81-096c5b944fb7","Type":"ContainerDied","Data":"b4e387c2cb4c4559e891db9b90b741782c7c86e2fd7b613450976204f6fcdd36"} Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.739710 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-njnmd" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.739752 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4e387c2cb4c4559e891db9b90b741782c7c86e2fd7b613450976204f6fcdd36" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.893611 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x"] Dec 03 22:41:30 crc kubenswrapper[4715]: E1203 22:41:30.894250 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10aa3b67-88a4-4a5d-9d81-096c5b944fb7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.894267 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="10aa3b67-88a4-4a5d-9d81-096c5b944fb7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 22:41:30 crc kubenswrapper[4715]: E1203 22:41:30.894299 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" containerName="extract-utilities" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.894306 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" containerName="extract-utilities" Dec 03 22:41:30 crc kubenswrapper[4715]: E1203 22:41:30.894318 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" containerName="extract-content" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.894326 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" containerName="extract-content" Dec 03 22:41:30 crc kubenswrapper[4715]: E1203 22:41:30.894336 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" containerName="registry-server" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.894343 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" containerName="registry-server" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.894539 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="10aa3b67-88a4-4a5d-9d81-096c5b944fb7" containerName="nova-edpm-deployment-openstack-edpm-ipam" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.894555 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="16bdae06-b96e-49f9-afce-21ccf312277d" containerName="registry-server" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.895606 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.901908 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.902154 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.902308 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.902326 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.902326 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fjgbf" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.905405 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x"] Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.949207 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.949308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.949379 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.949678 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.949714 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.949739 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:30 crc kubenswrapper[4715]: I1203 22:41:30.949844 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lh27\" (UniqueName: \"kubernetes.io/projected/32684f56-1e44-4d82-b8d2-f1193ba615d2-kube-api-access-7lh27\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.052990 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.053058 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.053097 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.053224 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lh27\" (UniqueName: \"kubernetes.io/projected/32684f56-1e44-4d82-b8d2-f1193ba615d2-kube-api-access-7lh27\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.053327 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.053365 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.053456 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.059904 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.059935 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.060205 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.060646 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.065517 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.065862 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.077639 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lh27\" (UniqueName: \"kubernetes.io/projected/32684f56-1e44-4d82-b8d2-f1193ba615d2-kube-api-access-7lh27\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.223680 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:41:31 crc kubenswrapper[4715]: I1203 22:41:31.880072 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x"] Dec 03 22:41:32 crc kubenswrapper[4715]: I1203 22:41:32.764206 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" event={"ID":"32684f56-1e44-4d82-b8d2-f1193ba615d2","Type":"ContainerStarted","Data":"a7b0390e9fc8c32c25746d654383a2d0262d7179483f56671a20829a8d446d32"} Dec 03 22:41:32 crc kubenswrapper[4715]: I1203 22:41:32.764654 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" event={"ID":"32684f56-1e44-4d82-b8d2-f1193ba615d2","Type":"ContainerStarted","Data":"9dcae3fbd5f014a5e48b3729efd21fa6c8b05c464a15f84164bce265d4670607"} Dec 03 22:41:32 crc kubenswrapper[4715]: I1203 22:41:32.793276 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" podStartSLOduration=2.356591805 podStartE2EDuration="2.793251736s" podCreationTimestamp="2025-12-03 22:41:30 +0000 UTC" firstStartedPulling="2025-12-03 22:41:31.885126974 +0000 UTC m=+3588.627837569" lastFinishedPulling="2025-12-03 22:41:32.321786905 +0000 UTC m=+3589.064497500" observedRunningTime="2025-12-03 22:41:32.786635497 +0000 UTC m=+3589.529346112" watchObservedRunningTime="2025-12-03 22:41:32.793251736 +0000 UTC m=+3589.535962341" Dec 03 22:41:35 crc kubenswrapper[4715]: I1203 22:41:35.159897 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:41:35 crc kubenswrapper[4715]: I1203 22:41:35.160687 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:42:05 crc kubenswrapper[4715]: I1203 22:42:05.160593 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:42:05 crc kubenswrapper[4715]: I1203 22:42:05.162830 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:42:05 crc kubenswrapper[4715]: I1203 22:42:05.162984 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:42:05 crc kubenswrapper[4715]: I1203 22:42:05.163937 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:42:05 crc kubenswrapper[4715]: I1203 22:42:05.164146 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" gracePeriod=600 Dec 03 22:42:05 crc kubenswrapper[4715]: E1203 22:42:05.794138 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:42:06 crc kubenswrapper[4715]: I1203 22:42:06.136653 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" exitCode=0 Dec 03 22:42:06 crc kubenswrapper[4715]: I1203 22:42:06.136696 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe"} Dec 03 22:42:06 crc kubenswrapper[4715]: I1203 22:42:06.136729 4715 scope.go:117] "RemoveContainer" containerID="7d9bfa2c1d02572cf252e696348da9eec449ec747bd736a0318c44de8c192b00" Dec 03 22:42:06 crc kubenswrapper[4715]: I1203 22:42:06.137552 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:42:06 crc kubenswrapper[4715]: E1203 22:42:06.138426 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:42:17 crc kubenswrapper[4715]: I1203 22:42:17.635813 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:42:17 crc kubenswrapper[4715]: E1203 22:42:17.637206 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:42:30 crc kubenswrapper[4715]: I1203 22:42:30.634788 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:42:30 crc kubenswrapper[4715]: E1203 22:42:30.635803 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:42:42 crc kubenswrapper[4715]: I1203 22:42:42.635123 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:42:42 crc kubenswrapper[4715]: E1203 22:42:42.637779 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:42:57 crc kubenswrapper[4715]: I1203 22:42:57.638360 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:42:57 crc kubenswrapper[4715]: E1203 22:42:57.642041 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:43:10 crc kubenswrapper[4715]: I1203 22:43:10.634288 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:43:10 crc kubenswrapper[4715]: E1203 22:43:10.634923 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:43:23 crc kubenswrapper[4715]: I1203 22:43:23.644827 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:43:23 crc kubenswrapper[4715]: E1203 22:43:23.645964 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:43:36 crc kubenswrapper[4715]: I1203 22:43:36.635579 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:43:36 crc kubenswrapper[4715]: E1203 22:43:36.636484 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:43:45 crc kubenswrapper[4715]: I1203 22:43:45.153701 4715 generic.go:334] "Generic (PLEG): container finished" podID="32684f56-1e44-4d82-b8d2-f1193ba615d2" containerID="a7b0390e9fc8c32c25746d654383a2d0262d7179483f56671a20829a8d446d32" exitCode=0 Dec 03 22:43:45 crc kubenswrapper[4715]: I1203 22:43:45.153821 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" event={"ID":"32684f56-1e44-4d82-b8d2-f1193ba615d2","Type":"ContainerDied","Data":"a7b0390e9fc8c32c25746d654383a2d0262d7179483f56671a20829a8d446d32"} Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.572452 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.647771 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ssh-key\") pod \"32684f56-1e44-4d82-b8d2-f1193ba615d2\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.647828 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lh27\" (UniqueName: \"kubernetes.io/projected/32684f56-1e44-4d82-b8d2-f1193ba615d2-kube-api-access-7lh27\") pod \"32684f56-1e44-4d82-b8d2-f1193ba615d2\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.647900 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-2\") pod \"32684f56-1e44-4d82-b8d2-f1193ba615d2\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.647935 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-telemetry-combined-ca-bundle\") pod \"32684f56-1e44-4d82-b8d2-f1193ba615d2\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.648009 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-1\") pod \"32684f56-1e44-4d82-b8d2-f1193ba615d2\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.648031 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-inventory\") pod \"32684f56-1e44-4d82-b8d2-f1193ba615d2\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.648050 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-0\") pod \"32684f56-1e44-4d82-b8d2-f1193ba615d2\" (UID: \"32684f56-1e44-4d82-b8d2-f1193ba615d2\") " Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.653671 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32684f56-1e44-4d82-b8d2-f1193ba615d2-kube-api-access-7lh27" (OuterVolumeSpecName: "kube-api-access-7lh27") pod "32684f56-1e44-4d82-b8d2-f1193ba615d2" (UID: "32684f56-1e44-4d82-b8d2-f1193ba615d2"). InnerVolumeSpecName "kube-api-access-7lh27". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.653844 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "32684f56-1e44-4d82-b8d2-f1193ba615d2" (UID: "32684f56-1e44-4d82-b8d2-f1193ba615d2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.681217 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "32684f56-1e44-4d82-b8d2-f1193ba615d2" (UID: "32684f56-1e44-4d82-b8d2-f1193ba615d2"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.681927 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "32684f56-1e44-4d82-b8d2-f1193ba615d2" (UID: "32684f56-1e44-4d82-b8d2-f1193ba615d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.682545 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "32684f56-1e44-4d82-b8d2-f1193ba615d2" (UID: "32684f56-1e44-4d82-b8d2-f1193ba615d2"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.683032 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "32684f56-1e44-4d82-b8d2-f1193ba615d2" (UID: "32684f56-1e44-4d82-b8d2-f1193ba615d2"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.683673 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-inventory" (OuterVolumeSpecName: "inventory") pod "32684f56-1e44-4d82-b8d2-f1193ba615d2" (UID: "32684f56-1e44-4d82-b8d2-f1193ba615d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.751235 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.751290 4715 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.751303 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.751312 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.751322 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lh27\" (UniqueName: \"kubernetes.io/projected/32684f56-1e44-4d82-b8d2-f1193ba615d2-kube-api-access-7lh27\") on node \"crc\" DevicePath \"\"" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.751330 4715 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 22:43:46 crc kubenswrapper[4715]: I1203 22:43:46.751339 4715 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/32684f56-1e44-4d82-b8d2-f1193ba615d2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 22:43:47 crc kubenswrapper[4715]: I1203 22:43:47.174494 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" event={"ID":"32684f56-1e44-4d82-b8d2-f1193ba615d2","Type":"ContainerDied","Data":"9dcae3fbd5f014a5e48b3729efd21fa6c8b05c464a15f84164bce265d4670607"} Dec 03 22:43:47 crc kubenswrapper[4715]: I1203 22:43:47.174820 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9dcae3fbd5f014a5e48b3729efd21fa6c8b05c464a15f84164bce265d4670607" Dec 03 22:43:47 crc kubenswrapper[4715]: I1203 22:43:47.174635 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x" Dec 03 22:43:49 crc kubenswrapper[4715]: I1203 22:43:49.634552 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:43:49 crc kubenswrapper[4715]: E1203 22:43:49.635047 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:44:02 crc kubenswrapper[4715]: I1203 22:44:02.635294 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:44:02 crc kubenswrapper[4715]: E1203 22:44:02.636286 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:44:16 crc kubenswrapper[4715]: I1203 22:44:16.634380 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:44:16 crc kubenswrapper[4715]: E1203 22:44:16.635358 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:44:27 crc kubenswrapper[4715]: I1203 22:44:27.634970 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:44:27 crc kubenswrapper[4715]: E1203 22:44:27.636046 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:44:41 crc kubenswrapper[4715]: I1203 22:44:41.634968 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:44:41 crc kubenswrapper[4715]: E1203 22:44:41.635897 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:44:53 crc kubenswrapper[4715]: I1203 22:44:53.640769 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:44:53 crc kubenswrapper[4715]: E1203 22:44:53.642993 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.164583 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9"] Dec 03 22:45:00 crc kubenswrapper[4715]: E1203 22:45:00.165435 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32684f56-1e44-4d82-b8d2-f1193ba615d2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.165450 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="32684f56-1e44-4d82-b8d2-f1193ba615d2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.165698 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="32684f56-1e44-4d82-b8d2-f1193ba615d2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.166416 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.169290 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.169325 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.179178 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9"] Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.335745 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gfdz\" (UniqueName: \"kubernetes.io/projected/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-kube-api-access-5gfdz\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.335980 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-secret-volume\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.336256 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-config-volume\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.438819 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-secret-volume\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.438973 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-config-volume\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.439019 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gfdz\" (UniqueName: \"kubernetes.io/projected/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-kube-api-access-5gfdz\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.440141 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-config-volume\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.445434 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-secret-volume\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.459470 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gfdz\" (UniqueName: \"kubernetes.io/projected/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-kube-api-access-5gfdz\") pod \"collect-profiles-29413365-r7qh9\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.503285 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.811229 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9"] Dec 03 22:45:00 crc kubenswrapper[4715]: I1203 22:45:00.931325 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" event={"ID":"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a","Type":"ContainerStarted","Data":"c7e506a214930fce1b3a8a1932d333ed5d90e496be63fe26fc0d9a05e7b31d53"} Dec 03 22:45:01 crc kubenswrapper[4715]: I1203 22:45:01.943988 4715 generic.go:334] "Generic (PLEG): container finished" podID="e13a8c1e-fc37-488f-bf2a-f873bbd0be9a" containerID="e1051a094b1d0fd32455c49ee700b7386abaa821225cca6f196c2b5687739456" exitCode=0 Dec 03 22:45:01 crc kubenswrapper[4715]: I1203 22:45:01.944077 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" event={"ID":"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a","Type":"ContainerDied","Data":"e1051a094b1d0fd32455c49ee700b7386abaa821225cca6f196c2b5687739456"} Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.484161 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.485826 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.488761 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.488929 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.489159 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.489755 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-7f5d9" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.495814 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.588996 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.589083 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.589110 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.589287 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-config-data\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.589430 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pvhs\" (UniqueName: \"kubernetes.io/projected/3c5223ae-7341-4aa4-821b-ec5450e37ebc-kube-api-access-2pvhs\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.589559 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.589766 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.589804 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.590038 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.691954 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692069 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692097 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692163 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692225 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692395 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692538 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692576 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692619 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-config-data\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692669 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pvhs\" (UniqueName: \"kubernetes.io/projected/3c5223ae-7341-4aa4-821b-ec5450e37ebc-kube-api-access-2pvhs\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.692870 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.693240 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.694122 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.694363 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-config-data\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.700649 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.700816 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.701558 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.720864 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pvhs\" (UniqueName: \"kubernetes.io/projected/3c5223ae-7341-4aa4-821b-ec5450e37ebc-kube-api-access-2pvhs\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.722800 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " pod="openstack/tempest-tests-tempest" Dec 03 22:45:02 crc kubenswrapper[4715]: I1203 22:45:02.814741 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.277815 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.283078 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.288977 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.406443 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-config-volume\") pod \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.406490 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-secret-volume\") pod \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.406565 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gfdz\" (UniqueName: \"kubernetes.io/projected/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-kube-api-access-5gfdz\") pod \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\" (UID: \"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a\") " Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.407657 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-config-volume" (OuterVolumeSpecName: "config-volume") pod "e13a8c1e-fc37-488f-bf2a-f873bbd0be9a" (UID: "e13a8c1e-fc37-488f-bf2a-f873bbd0be9a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.411152 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e13a8c1e-fc37-488f-bf2a-f873bbd0be9a" (UID: "e13a8c1e-fc37-488f-bf2a-f873bbd0be9a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.411252 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-kube-api-access-5gfdz" (OuterVolumeSpecName: "kube-api-access-5gfdz") pod "e13a8c1e-fc37-488f-bf2a-f873bbd0be9a" (UID: "e13a8c1e-fc37-488f-bf2a-f873bbd0be9a"). InnerVolumeSpecName "kube-api-access-5gfdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.509144 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.509231 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.509252 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gfdz\" (UniqueName: \"kubernetes.io/projected/e13a8c1e-fc37-488f-bf2a-f873bbd0be9a-kube-api-access-5gfdz\") on node \"crc\" DevicePath \"\"" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.968154 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3c5223ae-7341-4aa4-821b-ec5450e37ebc","Type":"ContainerStarted","Data":"51b3247ed5534bf356fa4fb44544509b08c524ffb025382b89d0a0068757573a"} Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.971915 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" event={"ID":"e13a8c1e-fc37-488f-bf2a-f873bbd0be9a","Type":"ContainerDied","Data":"c7e506a214930fce1b3a8a1932d333ed5d90e496be63fe26fc0d9a05e7b31d53"} Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.971954 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c7e506a214930fce1b3a8a1932d333ed5d90e496be63fe26fc0d9a05e7b31d53" Dec 03 22:45:03 crc kubenswrapper[4715]: I1203 22:45:03.972010 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413365-r7qh9" Dec 03 22:45:04 crc kubenswrapper[4715]: I1203 22:45:04.353685 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc"] Dec 03 22:45:04 crc kubenswrapper[4715]: I1203 22:45:04.363585 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413320-49jkc"] Dec 03 22:45:04 crc kubenswrapper[4715]: I1203 22:45:04.634019 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:45:04 crc kubenswrapper[4715]: E1203 22:45:04.634398 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:45:05 crc kubenswrapper[4715]: I1203 22:45:05.650155 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f97b6863-0677-43be-9fcd-d6c18bed4f90" path="/var/lib/kubelet/pods/f97b6863-0677-43be-9fcd-d6c18bed4f90/volumes" Dec 03 22:45:18 crc kubenswrapper[4715]: I1203 22:45:18.634119 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:45:18 crc kubenswrapper[4715]: E1203 22:45:18.634854 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:45:31 crc kubenswrapper[4715]: I1203 22:45:31.634609 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:45:31 crc kubenswrapper[4715]: E1203 22:45:31.635213 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:45:39 crc kubenswrapper[4715]: E1203 22:45:39.178242 4715 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 03 22:45:39 crc kubenswrapper[4715]: E1203 22:45:39.179255 4715 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2pvhs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(3c5223ae-7341-4aa4-821b-ec5450e37ebc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 22:45:39 crc kubenswrapper[4715]: E1203 22:45:39.180537 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="3c5223ae-7341-4aa4-821b-ec5450e37ebc" Dec 03 22:45:39 crc kubenswrapper[4715]: E1203 22:45:39.307216 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="3c5223ae-7341-4aa4-821b-ec5450e37ebc" Dec 03 22:45:42 crc kubenswrapper[4715]: I1203 22:45:42.634170 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:45:42 crc kubenswrapper[4715]: E1203 22:45:42.634631 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:45:54 crc kubenswrapper[4715]: I1203 22:45:54.113803 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 22:45:55 crc kubenswrapper[4715]: I1203 22:45:55.453387 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3c5223ae-7341-4aa4-821b-ec5450e37ebc","Type":"ContainerStarted","Data":"0b0e231b4437de350c57aa874ba905e6030ee222655536f4c8df0d44e621428a"} Dec 03 22:45:55 crc kubenswrapper[4715]: I1203 22:45:55.952303 4715 scope.go:117] "RemoveContainer" containerID="c2af3a574582e17f741374ded3450b91b586e46fa7b6cdd65812b502889e1a68" Dec 03 22:45:56 crc kubenswrapper[4715]: I1203 22:45:56.635101 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:45:56 crc kubenswrapper[4715]: E1203 22:45:56.635623 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:46:07 crc kubenswrapper[4715]: I1203 22:46:07.634749 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:46:07 crc kubenswrapper[4715]: E1203 22:46:07.635455 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:46:19 crc kubenswrapper[4715]: I1203 22:46:19.637533 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:46:19 crc kubenswrapper[4715]: E1203 22:46:19.638906 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:46:32 crc kubenswrapper[4715]: I1203 22:46:32.634835 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:46:32 crc kubenswrapper[4715]: E1203 22:46:32.635789 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:46:47 crc kubenswrapper[4715]: I1203 22:46:47.635040 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:46:47 crc kubenswrapper[4715]: E1203 22:46:47.635856 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:46:58 crc kubenswrapper[4715]: I1203 22:46:58.633834 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:46:58 crc kubenswrapper[4715]: E1203 22:46:58.634779 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:47:12 crc kubenswrapper[4715]: I1203 22:47:12.634745 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:47:13 crc kubenswrapper[4715]: I1203 22:47:13.189264 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"f37b824df953b26abbc5aaf2b2af398da5e300d4610032a787f3edd67e7adf4e"} Dec 03 22:47:13 crc kubenswrapper[4715]: I1203 22:47:13.220289 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=81.391968958 podStartE2EDuration="2m12.220263262s" podCreationTimestamp="2025-12-03 22:45:01 +0000 UTC" firstStartedPulling="2025-12-03 22:45:03.282888884 +0000 UTC m=+3800.025599479" lastFinishedPulling="2025-12-03 22:45:54.111183178 +0000 UTC m=+3850.853893783" observedRunningTime="2025-12-03 22:45:55.475402761 +0000 UTC m=+3852.218113376" watchObservedRunningTime="2025-12-03 22:47:13.220263262 +0000 UTC m=+3929.962973867" Dec 03 22:49:35 crc kubenswrapper[4715]: I1203 22:49:35.160127 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:49:35 crc kubenswrapper[4715]: I1203 22:49:35.161813 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.438625 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mqrzj"] Dec 03 22:49:47 crc kubenswrapper[4715]: E1203 22:49:47.439642 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e13a8c1e-fc37-488f-bf2a-f873bbd0be9a" containerName="collect-profiles" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.439660 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e13a8c1e-fc37-488f-bf2a-f873bbd0be9a" containerName="collect-profiles" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.439891 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e13a8c1e-fc37-488f-bf2a-f873bbd0be9a" containerName="collect-profiles" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.441832 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.460443 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqrzj"] Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.581391 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-catalog-content\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.581464 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t4pk\" (UniqueName: \"kubernetes.io/projected/74b49e59-e408-4ef1-a625-eeb5a73b29bb-kube-api-access-4t4pk\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.581568 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-utilities\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.683099 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-catalog-content\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.683183 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t4pk\" (UniqueName: \"kubernetes.io/projected/74b49e59-e408-4ef1-a625-eeb5a73b29bb-kube-api-access-4t4pk\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.683213 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-utilities\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.683784 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-utilities\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:47 crc kubenswrapper[4715]: I1203 22:49:47.684035 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-catalog-content\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:48 crc kubenswrapper[4715]: I1203 22:49:48.026409 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t4pk\" (UniqueName: \"kubernetes.io/projected/74b49e59-e408-4ef1-a625-eeb5a73b29bb-kube-api-access-4t4pk\") pod \"certified-operators-mqrzj\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:48 crc kubenswrapper[4715]: I1203 22:49:48.063876 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:48 crc kubenswrapper[4715]: I1203 22:49:48.497040 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mqrzj"] Dec 03 22:49:48 crc kubenswrapper[4715]: I1203 22:49:48.720366 4715 generic.go:334] "Generic (PLEG): container finished" podID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerID="d3ba6423c0961fc2e64fefc132c4efd224d3dab3f4cd51286c1bb87ac1628cdd" exitCode=0 Dec 03 22:49:48 crc kubenswrapper[4715]: I1203 22:49:48.720417 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqrzj" event={"ID":"74b49e59-e408-4ef1-a625-eeb5a73b29bb","Type":"ContainerDied","Data":"d3ba6423c0961fc2e64fefc132c4efd224d3dab3f4cd51286c1bb87ac1628cdd"} Dec 03 22:49:48 crc kubenswrapper[4715]: I1203 22:49:48.720446 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqrzj" event={"ID":"74b49e59-e408-4ef1-a625-eeb5a73b29bb","Type":"ContainerStarted","Data":"ee025c8320b16dd89ab45d14c128280bff6b881691fd4b6002c60ce10ee71572"} Dec 03 22:49:49 crc kubenswrapper[4715]: I1203 22:49:49.738296 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqrzj" event={"ID":"74b49e59-e408-4ef1-a625-eeb5a73b29bb","Type":"ContainerStarted","Data":"abf068aad8ca7c6f1fcf055463eb7680ea3b5a5d5e989d5d36df749f0694ad75"} Dec 03 22:49:50 crc kubenswrapper[4715]: I1203 22:49:50.749588 4715 generic.go:334] "Generic (PLEG): container finished" podID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerID="abf068aad8ca7c6f1fcf055463eb7680ea3b5a5d5e989d5d36df749f0694ad75" exitCode=0 Dec 03 22:49:50 crc kubenswrapper[4715]: I1203 22:49:50.749651 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqrzj" event={"ID":"74b49e59-e408-4ef1-a625-eeb5a73b29bb","Type":"ContainerDied","Data":"abf068aad8ca7c6f1fcf055463eb7680ea3b5a5d5e989d5d36df749f0694ad75"} Dec 03 22:49:51 crc kubenswrapper[4715]: I1203 22:49:51.760938 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqrzj" event={"ID":"74b49e59-e408-4ef1-a625-eeb5a73b29bb","Type":"ContainerStarted","Data":"e3231676ed3fa1f6d9425316a6810e5fb432711273af96d8c6ecc5f6326d52be"} Dec 03 22:49:51 crc kubenswrapper[4715]: I1203 22:49:51.786722 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mqrzj" podStartSLOduration=2.28842817 podStartE2EDuration="4.786703617s" podCreationTimestamp="2025-12-03 22:49:47 +0000 UTC" firstStartedPulling="2025-12-03 22:49:48.723015345 +0000 UTC m=+4085.465725940" lastFinishedPulling="2025-12-03 22:49:51.221290792 +0000 UTC m=+4087.964001387" observedRunningTime="2025-12-03 22:49:51.776407289 +0000 UTC m=+4088.519117894" watchObservedRunningTime="2025-12-03 22:49:51.786703617 +0000 UTC m=+4088.529414212" Dec 03 22:49:58 crc kubenswrapper[4715]: I1203 22:49:58.064199 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:58 crc kubenswrapper[4715]: I1203 22:49:58.064679 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:58 crc kubenswrapper[4715]: I1203 22:49:58.113296 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:58 crc kubenswrapper[4715]: I1203 22:49:58.888926 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:49:58 crc kubenswrapper[4715]: I1203 22:49:58.954941 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqrzj"] Dec 03 22:50:00 crc kubenswrapper[4715]: I1203 22:50:00.847943 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mqrzj" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerName="registry-server" containerID="cri-o://e3231676ed3fa1f6d9425316a6810e5fb432711273af96d8c6ecc5f6326d52be" gracePeriod=2 Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.861867 4715 generic.go:334] "Generic (PLEG): container finished" podID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerID="e3231676ed3fa1f6d9425316a6810e5fb432711273af96d8c6ecc5f6326d52be" exitCode=0 Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.861994 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqrzj" event={"ID":"74b49e59-e408-4ef1-a625-eeb5a73b29bb","Type":"ContainerDied","Data":"e3231676ed3fa1f6d9425316a6810e5fb432711273af96d8c6ecc5f6326d52be"} Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.862163 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mqrzj" event={"ID":"74b49e59-e408-4ef1-a625-eeb5a73b29bb","Type":"ContainerDied","Data":"ee025c8320b16dd89ab45d14c128280bff6b881691fd4b6002c60ce10ee71572"} Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.862183 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee025c8320b16dd89ab45d14c128280bff6b881691fd4b6002c60ce10ee71572" Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.902026 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.997442 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-catalog-content\") pod \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.997604 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-utilities\") pod \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.997684 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t4pk\" (UniqueName: \"kubernetes.io/projected/74b49e59-e408-4ef1-a625-eeb5a73b29bb-kube-api-access-4t4pk\") pod \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\" (UID: \"74b49e59-e408-4ef1-a625-eeb5a73b29bb\") " Dec 03 22:50:01 crc kubenswrapper[4715]: I1203 22:50:01.998870 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-utilities" (OuterVolumeSpecName: "utilities") pod "74b49e59-e408-4ef1-a625-eeb5a73b29bb" (UID: "74b49e59-e408-4ef1-a625-eeb5a73b29bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:50:02 crc kubenswrapper[4715]: I1203 22:50:02.004782 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74b49e59-e408-4ef1-a625-eeb5a73b29bb-kube-api-access-4t4pk" (OuterVolumeSpecName: "kube-api-access-4t4pk") pod "74b49e59-e408-4ef1-a625-eeb5a73b29bb" (UID: "74b49e59-e408-4ef1-a625-eeb5a73b29bb"). InnerVolumeSpecName "kube-api-access-4t4pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:50:02 crc kubenswrapper[4715]: I1203 22:50:02.048314 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74b49e59-e408-4ef1-a625-eeb5a73b29bb" (UID: "74b49e59-e408-4ef1-a625-eeb5a73b29bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:50:02 crc kubenswrapper[4715]: I1203 22:50:02.099789 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:02 crc kubenswrapper[4715]: I1203 22:50:02.100052 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74b49e59-e408-4ef1-a625-eeb5a73b29bb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:02 crc kubenswrapper[4715]: I1203 22:50:02.100130 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t4pk\" (UniqueName: \"kubernetes.io/projected/74b49e59-e408-4ef1-a625-eeb5a73b29bb-kube-api-access-4t4pk\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:02 crc kubenswrapper[4715]: I1203 22:50:02.870763 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mqrzj" Dec 03 22:50:02 crc kubenswrapper[4715]: I1203 22:50:02.913983 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mqrzj"] Dec 03 22:50:02 crc kubenswrapper[4715]: I1203 22:50:02.923352 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mqrzj"] Dec 03 22:50:03 crc kubenswrapper[4715]: E1203 22:50:03.067032 4715 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74b49e59_e408_4ef1_a625_eeb5a73b29bb.slice\": RecentStats: unable to find data in memory cache]" Dec 03 22:50:03 crc kubenswrapper[4715]: I1203 22:50:03.653233 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" path="/var/lib/kubelet/pods/74b49e59-e408-4ef1-a625-eeb5a73b29bb/volumes" Dec 03 22:50:05 crc kubenswrapper[4715]: I1203 22:50:05.159545 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:50:05 crc kubenswrapper[4715]: I1203 22:50:05.159606 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.005451 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7wgft"] Dec 03 22:50:08 crc kubenswrapper[4715]: E1203 22:50:08.006293 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerName="extract-content" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.006309 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerName="extract-content" Dec 03 22:50:08 crc kubenswrapper[4715]: E1203 22:50:08.006332 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerName="registry-server" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.006339 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerName="registry-server" Dec 03 22:50:08 crc kubenswrapper[4715]: E1203 22:50:08.006384 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerName="extract-utilities" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.006392 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerName="extract-utilities" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.006622 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="74b49e59-e408-4ef1-a625-eeb5a73b29bb" containerName="registry-server" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.008393 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.021557 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7wgft"] Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.108898 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sjm2\" (UniqueName: \"kubernetes.io/projected/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-kube-api-access-8sjm2\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.109138 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-catalog-content\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.109298 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-utilities\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.211522 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sjm2\" (UniqueName: \"kubernetes.io/projected/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-kube-api-access-8sjm2\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.211604 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-catalog-content\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.211631 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-utilities\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.212459 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-utilities\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.212569 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-catalog-content\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.230192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sjm2\" (UniqueName: \"kubernetes.io/projected/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-kube-api-access-8sjm2\") pod \"redhat-marketplace-7wgft\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.349999 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.797790 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7wgft"] Dec 03 22:50:08 crc kubenswrapper[4715]: W1203 22:50:08.803988 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccfdf1b4_0b40_4ec0_9cc2_c6dffbc60352.slice/crio-46970cb395b23766ced32f85acfd0eea84c33fbfed5868f606edd7ec6daea860 WatchSource:0}: Error finding container 46970cb395b23766ced32f85acfd0eea84c33fbfed5868f606edd7ec6daea860: Status 404 returned error can't find the container with id 46970cb395b23766ced32f85acfd0eea84c33fbfed5868f606edd7ec6daea860 Dec 03 22:50:08 crc kubenswrapper[4715]: I1203 22:50:08.936815 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7wgft" event={"ID":"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352","Type":"ContainerStarted","Data":"46970cb395b23766ced32f85acfd0eea84c33fbfed5868f606edd7ec6daea860"} Dec 03 22:50:09 crc kubenswrapper[4715]: I1203 22:50:09.948815 4715 generic.go:334] "Generic (PLEG): container finished" podID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerID="9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd" exitCode=0 Dec 03 22:50:09 crc kubenswrapper[4715]: I1203 22:50:09.948866 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7wgft" event={"ID":"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352","Type":"ContainerDied","Data":"9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd"} Dec 03 22:50:09 crc kubenswrapper[4715]: I1203 22:50:09.951327 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:50:10 crc kubenswrapper[4715]: I1203 22:50:10.960288 4715 generic.go:334] "Generic (PLEG): container finished" podID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerID="a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865" exitCode=0 Dec 03 22:50:10 crc kubenswrapper[4715]: I1203 22:50:10.960398 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7wgft" event={"ID":"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352","Type":"ContainerDied","Data":"a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865"} Dec 03 22:50:11 crc kubenswrapper[4715]: I1203 22:50:11.974096 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7wgft" event={"ID":"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352","Type":"ContainerStarted","Data":"b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc"} Dec 03 22:50:11 crc kubenswrapper[4715]: I1203 22:50:11.998072 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7wgft" podStartSLOduration=3.576842782 podStartE2EDuration="4.998047689s" podCreationTimestamp="2025-12-03 22:50:07 +0000 UTC" firstStartedPulling="2025-12-03 22:50:09.951073728 +0000 UTC m=+4106.693784343" lastFinishedPulling="2025-12-03 22:50:11.372278645 +0000 UTC m=+4108.114989250" observedRunningTime="2025-12-03 22:50:11.992629322 +0000 UTC m=+4108.735339937" watchObservedRunningTime="2025-12-03 22:50:11.998047689 +0000 UTC m=+4108.740758284" Dec 03 22:50:18 crc kubenswrapper[4715]: I1203 22:50:18.350880 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:18 crc kubenswrapper[4715]: I1203 22:50:18.351489 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:18 crc kubenswrapper[4715]: I1203 22:50:18.421381 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:19 crc kubenswrapper[4715]: I1203 22:50:19.095358 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:19 crc kubenswrapper[4715]: I1203 22:50:19.146190 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7wgft"] Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.064081 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7wgft" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerName="registry-server" containerID="cri-o://b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc" gracePeriod=2 Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.621448 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.785478 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-catalog-content\") pod \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.785920 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-utilities\") pod \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.786146 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sjm2\" (UniqueName: \"kubernetes.io/projected/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-kube-api-access-8sjm2\") pod \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\" (UID: \"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352\") " Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.787620 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-utilities" (OuterVolumeSpecName: "utilities") pod "ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" (UID: "ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.792483 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-kube-api-access-8sjm2" (OuterVolumeSpecName: "kube-api-access-8sjm2") pod "ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" (UID: "ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352"). InnerVolumeSpecName "kube-api-access-8sjm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.810910 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" (UID: "ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.888591 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sjm2\" (UniqueName: \"kubernetes.io/projected/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-kube-api-access-8sjm2\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.888889 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:21 crc kubenswrapper[4715]: I1203 22:50:21.888950 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.077154 4715 generic.go:334] "Generic (PLEG): container finished" podID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerID="b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc" exitCode=0 Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.077223 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7wgft" event={"ID":"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352","Type":"ContainerDied","Data":"b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc"} Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.077274 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7wgft" event={"ID":"ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352","Type":"ContainerDied","Data":"46970cb395b23766ced32f85acfd0eea84c33fbfed5868f606edd7ec6daea860"} Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.077310 4715 scope.go:117] "RemoveContainer" containerID="b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.077597 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7wgft" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.107047 4715 scope.go:117] "RemoveContainer" containerID="a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.141102 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7wgft"] Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.149716 4715 scope.go:117] "RemoveContainer" containerID="9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.159665 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7wgft"] Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.188686 4715 scope.go:117] "RemoveContainer" containerID="b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc" Dec 03 22:50:22 crc kubenswrapper[4715]: E1203 22:50:22.189325 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc\": container with ID starting with b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc not found: ID does not exist" containerID="b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.189404 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc"} err="failed to get container status \"b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc\": rpc error: code = NotFound desc = could not find container \"b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc\": container with ID starting with b10e7d684450b3abe421fabcee3b035c7dca320352788a44423307247b49dddc not found: ID does not exist" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.189467 4715 scope.go:117] "RemoveContainer" containerID="a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865" Dec 03 22:50:22 crc kubenswrapper[4715]: E1203 22:50:22.189951 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865\": container with ID starting with a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865 not found: ID does not exist" containerID="a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.190014 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865"} err="failed to get container status \"a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865\": rpc error: code = NotFound desc = could not find container \"a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865\": container with ID starting with a9a2dc76fea30addedf216722e5cdc335964e1433cc3c0e1d96154fdbb243865 not found: ID does not exist" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.190048 4715 scope.go:117] "RemoveContainer" containerID="9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd" Dec 03 22:50:22 crc kubenswrapper[4715]: E1203 22:50:22.190464 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd\": container with ID starting with 9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd not found: ID does not exist" containerID="9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd" Dec 03 22:50:22 crc kubenswrapper[4715]: I1203 22:50:22.190492 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd"} err="failed to get container status \"9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd\": rpc error: code = NotFound desc = could not find container \"9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd\": container with ID starting with 9de21a7d013e5dd63138987878594944a6f3df19073f8f568d094d46625ae0cd not found: ID does not exist" Dec 03 22:50:23 crc kubenswrapper[4715]: I1203 22:50:23.644675 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" path="/var/lib/kubelet/pods/ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352/volumes" Dec 03 22:50:35 crc kubenswrapper[4715]: I1203 22:50:35.160642 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:50:35 crc kubenswrapper[4715]: I1203 22:50:35.161356 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:50:35 crc kubenswrapper[4715]: I1203 22:50:35.161423 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:50:35 crc kubenswrapper[4715]: I1203 22:50:35.162475 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f37b824df953b26abbc5aaf2b2af398da5e300d4610032a787f3edd67e7adf4e"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:50:35 crc kubenswrapper[4715]: I1203 22:50:35.162613 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://f37b824df953b26abbc5aaf2b2af398da5e300d4610032a787f3edd67e7adf4e" gracePeriod=600 Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.215606 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4r4sv"] Dec 03 22:50:36 crc kubenswrapper[4715]: E1203 22:50:36.216467 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerName="extract-utilities" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.216482 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerName="extract-utilities" Dec 03 22:50:36 crc kubenswrapper[4715]: E1203 22:50:36.216571 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerName="extract-content" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.216578 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerName="extract-content" Dec 03 22:50:36 crc kubenswrapper[4715]: E1203 22:50:36.216588 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerName="registry-server" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.216593 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerName="registry-server" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.216796 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccfdf1b4-0b40-4ec0-9cc2-c6dffbc60352" containerName="registry-server" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.218277 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.221240 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="f37b824df953b26abbc5aaf2b2af398da5e300d4610032a787f3edd67e7adf4e" exitCode=0 Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.221291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"f37b824df953b26abbc5aaf2b2af398da5e300d4610032a787f3edd67e7adf4e"} Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.221324 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828"} Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.221343 4715 scope.go:117] "RemoveContainer" containerID="c0364fdbbe29c5f38cec3a2cccf5852bd62b59d3c4829d9b8b71689410d91cbe" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.229141 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4r4sv"] Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.279119 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-catalog-content\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.279338 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-utilities\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.279702 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9qm9\" (UniqueName: \"kubernetes.io/projected/87d892b8-24ea-4e58-ac8a-2ed88e721791-kube-api-access-h9qm9\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.381461 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-catalog-content\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.381585 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-utilities\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.381628 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9qm9\" (UniqueName: \"kubernetes.io/projected/87d892b8-24ea-4e58-ac8a-2ed88e721791-kube-api-access-h9qm9\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.381990 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-utilities\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.382050 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-catalog-content\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.400424 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9qm9\" (UniqueName: \"kubernetes.io/projected/87d892b8-24ea-4e58-ac8a-2ed88e721791-kube-api-access-h9qm9\") pod \"community-operators-4r4sv\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:36 crc kubenswrapper[4715]: I1203 22:50:36.555081 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:37 crc kubenswrapper[4715]: I1203 22:50:37.250825 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4r4sv"] Dec 03 22:50:38 crc kubenswrapper[4715]: I1203 22:50:38.252055 4715 generic.go:334] "Generic (PLEG): container finished" podID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerID="a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef" exitCode=0 Dec 03 22:50:38 crc kubenswrapper[4715]: I1203 22:50:38.253185 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4r4sv" event={"ID":"87d892b8-24ea-4e58-ac8a-2ed88e721791","Type":"ContainerDied","Data":"a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef"} Dec 03 22:50:38 crc kubenswrapper[4715]: I1203 22:50:38.255609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4r4sv" event={"ID":"87d892b8-24ea-4e58-ac8a-2ed88e721791","Type":"ContainerStarted","Data":"f91e31bc5ed00eb2b44101ff409281d37003cec6f565f67bb38686d3852cf401"} Dec 03 22:50:40 crc kubenswrapper[4715]: I1203 22:50:40.273975 4715 generic.go:334] "Generic (PLEG): container finished" podID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerID="b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb" exitCode=0 Dec 03 22:50:40 crc kubenswrapper[4715]: I1203 22:50:40.274055 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4r4sv" event={"ID":"87d892b8-24ea-4e58-ac8a-2ed88e721791","Type":"ContainerDied","Data":"b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb"} Dec 03 22:50:42 crc kubenswrapper[4715]: I1203 22:50:42.292126 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4r4sv" event={"ID":"87d892b8-24ea-4e58-ac8a-2ed88e721791","Type":"ContainerStarted","Data":"a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8"} Dec 03 22:50:42 crc kubenswrapper[4715]: I1203 22:50:42.309421 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4r4sv" podStartSLOduration=3.797042122 podStartE2EDuration="6.309402559s" podCreationTimestamp="2025-12-03 22:50:36 +0000 UTC" firstStartedPulling="2025-12-03 22:50:38.254988816 +0000 UTC m=+4134.997699451" lastFinishedPulling="2025-12-03 22:50:40.767349293 +0000 UTC m=+4137.510059888" observedRunningTime="2025-12-03 22:50:42.307747985 +0000 UTC m=+4139.050458590" watchObservedRunningTime="2025-12-03 22:50:42.309402559 +0000 UTC m=+4139.052113154" Dec 03 22:50:46 crc kubenswrapper[4715]: I1203 22:50:46.556204 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:46 crc kubenswrapper[4715]: I1203 22:50:46.556939 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:46 crc kubenswrapper[4715]: I1203 22:50:46.627326 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:47 crc kubenswrapper[4715]: I1203 22:50:47.432292 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:47 crc kubenswrapper[4715]: I1203 22:50:47.501980 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4r4sv"] Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.282605 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gmrmp"] Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.286563 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.294845 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gmrmp"] Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.383922 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4r4sv" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerName="registry-server" containerID="cri-o://a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8" gracePeriod=2 Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.450224 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q7hv\" (UniqueName: \"kubernetes.io/projected/f7c46fbd-c2df-492a-9a4b-4e511761edac-kube-api-access-8q7hv\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.450444 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-utilities\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.450549 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-catalog-content\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.551980 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q7hv\" (UniqueName: \"kubernetes.io/projected/f7c46fbd-c2df-492a-9a4b-4e511761edac-kube-api-access-8q7hv\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.552121 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-utilities\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.552173 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-catalog-content\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.552770 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-catalog-content\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.552894 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-utilities\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.573211 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q7hv\" (UniqueName: \"kubernetes.io/projected/f7c46fbd-c2df-492a-9a4b-4e511761edac-kube-api-access-8q7hv\") pod \"redhat-operators-gmrmp\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.618338 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.836802 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.960755 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-catalog-content\") pod \"87d892b8-24ea-4e58-ac8a-2ed88e721791\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.960917 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-utilities\") pod \"87d892b8-24ea-4e58-ac8a-2ed88e721791\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.961003 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9qm9\" (UniqueName: \"kubernetes.io/projected/87d892b8-24ea-4e58-ac8a-2ed88e721791-kube-api-access-h9qm9\") pod \"87d892b8-24ea-4e58-ac8a-2ed88e721791\" (UID: \"87d892b8-24ea-4e58-ac8a-2ed88e721791\") " Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.962130 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-utilities" (OuterVolumeSpecName: "utilities") pod "87d892b8-24ea-4e58-ac8a-2ed88e721791" (UID: "87d892b8-24ea-4e58-ac8a-2ed88e721791"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:50:49 crc kubenswrapper[4715]: I1203 22:50:49.966889 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87d892b8-24ea-4e58-ac8a-2ed88e721791-kube-api-access-h9qm9" (OuterVolumeSpecName: "kube-api-access-h9qm9") pod "87d892b8-24ea-4e58-ac8a-2ed88e721791" (UID: "87d892b8-24ea-4e58-ac8a-2ed88e721791"). InnerVolumeSpecName "kube-api-access-h9qm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.014609 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87d892b8-24ea-4e58-ac8a-2ed88e721791" (UID: "87d892b8-24ea-4e58-ac8a-2ed88e721791"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.063555 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.063586 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87d892b8-24ea-4e58-ac8a-2ed88e721791-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.063598 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9qm9\" (UniqueName: \"kubernetes.io/projected/87d892b8-24ea-4e58-ac8a-2ed88e721791-kube-api-access-h9qm9\") on node \"crc\" DevicePath \"\"" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.138115 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gmrmp"] Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.394800 4715 generic.go:334] "Generic (PLEG): container finished" podID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerID="66560681be74972d7a786fdb4a7e6464caaec5fe81c194980653d9a8d62de66f" exitCode=0 Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.394859 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gmrmp" event={"ID":"f7c46fbd-c2df-492a-9a4b-4e511761edac","Type":"ContainerDied","Data":"66560681be74972d7a786fdb4a7e6464caaec5fe81c194980653d9a8d62de66f"} Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.395157 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gmrmp" event={"ID":"f7c46fbd-c2df-492a-9a4b-4e511761edac","Type":"ContainerStarted","Data":"d8eea10a665cd997f7d6416e781a860b53dea1a546ee2d0406373066cad44ce5"} Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.397947 4715 generic.go:334] "Generic (PLEG): container finished" podID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerID="a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8" exitCode=0 Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.397993 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4r4sv" event={"ID":"87d892b8-24ea-4e58-ac8a-2ed88e721791","Type":"ContainerDied","Data":"a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8"} Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.398021 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4r4sv" event={"ID":"87d892b8-24ea-4e58-ac8a-2ed88e721791","Type":"ContainerDied","Data":"f91e31bc5ed00eb2b44101ff409281d37003cec6f565f67bb38686d3852cf401"} Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.398038 4715 scope.go:117] "RemoveContainer" containerID="a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.398171 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4r4sv" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.424623 4715 scope.go:117] "RemoveContainer" containerID="b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.428256 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4r4sv"] Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.435654 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4r4sv"] Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.445208 4715 scope.go:117] "RemoveContainer" containerID="a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.461991 4715 scope.go:117] "RemoveContainer" containerID="a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8" Dec 03 22:50:50 crc kubenswrapper[4715]: E1203 22:50:50.462486 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8\": container with ID starting with a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8 not found: ID does not exist" containerID="a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.462677 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8"} err="failed to get container status \"a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8\": rpc error: code = NotFound desc = could not find container \"a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8\": container with ID starting with a99c765a57d061699bc900382af137e80e89ab90fc00f4964dcba19886ab5be8 not found: ID does not exist" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.462725 4715 scope.go:117] "RemoveContainer" containerID="b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb" Dec 03 22:50:50 crc kubenswrapper[4715]: E1203 22:50:50.463143 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb\": container with ID starting with b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb not found: ID does not exist" containerID="b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.463188 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb"} err="failed to get container status \"b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb\": rpc error: code = NotFound desc = could not find container \"b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb\": container with ID starting with b7a6b255e722f1531fe8cf705fa76f94f05bfc1dfbb4f9b0178e23ea26ec3afb not found: ID does not exist" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.463201 4715 scope.go:117] "RemoveContainer" containerID="a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef" Dec 03 22:50:50 crc kubenswrapper[4715]: E1203 22:50:50.463471 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef\": container with ID starting with a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef not found: ID does not exist" containerID="a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef" Dec 03 22:50:50 crc kubenswrapper[4715]: I1203 22:50:50.463491 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef"} err="failed to get container status \"a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef\": rpc error: code = NotFound desc = could not find container \"a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef\": container with ID starting with a7dab0eaa0c971c3cfcfda747214528cc7e6ac2bec99ce09a2c2dcc0eb3b27ef not found: ID does not exist" Dec 03 22:50:51 crc kubenswrapper[4715]: I1203 22:50:51.643972 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" path="/var/lib/kubelet/pods/87d892b8-24ea-4e58-ac8a-2ed88e721791/volumes" Dec 03 22:50:52 crc kubenswrapper[4715]: I1203 22:50:52.434732 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gmrmp" event={"ID":"f7c46fbd-c2df-492a-9a4b-4e511761edac","Type":"ContainerStarted","Data":"fa0ecc8b2f0ddce0313119bd2bce817d8548fc218ae233aa56272fd2535494f7"} Dec 03 22:50:53 crc kubenswrapper[4715]: I1203 22:50:53.446783 4715 generic.go:334] "Generic (PLEG): container finished" podID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerID="fa0ecc8b2f0ddce0313119bd2bce817d8548fc218ae233aa56272fd2535494f7" exitCode=0 Dec 03 22:50:53 crc kubenswrapper[4715]: I1203 22:50:53.447058 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gmrmp" event={"ID":"f7c46fbd-c2df-492a-9a4b-4e511761edac","Type":"ContainerDied","Data":"fa0ecc8b2f0ddce0313119bd2bce817d8548fc218ae233aa56272fd2535494f7"} Dec 03 22:50:56 crc kubenswrapper[4715]: I1203 22:50:56.492129 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gmrmp" event={"ID":"f7c46fbd-c2df-492a-9a4b-4e511761edac","Type":"ContainerStarted","Data":"2849fba46a997fb2750dd28b2db32e16537e7cc8cfac1f9f7e9c15342c8e898e"} Dec 03 22:50:56 crc kubenswrapper[4715]: I1203 22:50:56.532218 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gmrmp" podStartSLOduration=2.598715087 podStartE2EDuration="7.532192019s" podCreationTimestamp="2025-12-03 22:50:49 +0000 UTC" firstStartedPulling="2025-12-03 22:50:50.39687373 +0000 UTC m=+4147.139584325" lastFinishedPulling="2025-12-03 22:50:55.330350652 +0000 UTC m=+4152.073061257" observedRunningTime="2025-12-03 22:50:56.513482144 +0000 UTC m=+4153.256192739" watchObservedRunningTime="2025-12-03 22:50:56.532192019 +0000 UTC m=+4153.274902644" Dec 03 22:50:59 crc kubenswrapper[4715]: I1203 22:50:59.618733 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:50:59 crc kubenswrapper[4715]: I1203 22:50:59.619292 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:51:00 crc kubenswrapper[4715]: I1203 22:51:00.664905 4715 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gmrmp" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="registry-server" probeResult="failure" output=< Dec 03 22:51:00 crc kubenswrapper[4715]: timeout: failed to connect service ":50051" within 1s Dec 03 22:51:00 crc kubenswrapper[4715]: > Dec 03 22:51:09 crc kubenswrapper[4715]: I1203 22:51:09.665093 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:51:09 crc kubenswrapper[4715]: I1203 22:51:09.725830 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:51:09 crc kubenswrapper[4715]: I1203 22:51:09.899365 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gmrmp"] Dec 03 22:51:11 crc kubenswrapper[4715]: I1203 22:51:11.628129 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gmrmp" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="registry-server" containerID="cri-o://2849fba46a997fb2750dd28b2db32e16537e7cc8cfac1f9f7e9c15342c8e898e" gracePeriod=2 Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.641592 4715 generic.go:334] "Generic (PLEG): container finished" podID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerID="2849fba46a997fb2750dd28b2db32e16537e7cc8cfac1f9f7e9c15342c8e898e" exitCode=0 Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.641660 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gmrmp" event={"ID":"f7c46fbd-c2df-492a-9a4b-4e511761edac","Type":"ContainerDied","Data":"2849fba46a997fb2750dd28b2db32e16537e7cc8cfac1f9f7e9c15342c8e898e"} Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.642831 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gmrmp" event={"ID":"f7c46fbd-c2df-492a-9a4b-4e511761edac","Type":"ContainerDied","Data":"d8eea10a665cd997f7d6416e781a860b53dea1a546ee2d0406373066cad44ce5"} Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.642863 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8eea10a665cd997f7d6416e781a860b53dea1a546ee2d0406373066cad44ce5" Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.678565 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.809268 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8q7hv\" (UniqueName: \"kubernetes.io/projected/f7c46fbd-c2df-492a-9a4b-4e511761edac-kube-api-access-8q7hv\") pod \"f7c46fbd-c2df-492a-9a4b-4e511761edac\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.809676 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-catalog-content\") pod \"f7c46fbd-c2df-492a-9a4b-4e511761edac\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.809901 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-utilities\") pod \"f7c46fbd-c2df-492a-9a4b-4e511761edac\" (UID: \"f7c46fbd-c2df-492a-9a4b-4e511761edac\") " Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.810758 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-utilities" (OuterVolumeSpecName: "utilities") pod "f7c46fbd-c2df-492a-9a4b-4e511761edac" (UID: "f7c46fbd-c2df-492a-9a4b-4e511761edac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.811317 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.815765 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7c46fbd-c2df-492a-9a4b-4e511761edac-kube-api-access-8q7hv" (OuterVolumeSpecName: "kube-api-access-8q7hv") pod "f7c46fbd-c2df-492a-9a4b-4e511761edac" (UID: "f7c46fbd-c2df-492a-9a4b-4e511761edac"). InnerVolumeSpecName "kube-api-access-8q7hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.913617 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8q7hv\" (UniqueName: \"kubernetes.io/projected/f7c46fbd-c2df-492a-9a4b-4e511761edac-kube-api-access-8q7hv\") on node \"crc\" DevicePath \"\"" Dec 03 22:51:12 crc kubenswrapper[4715]: I1203 22:51:12.918909 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7c46fbd-c2df-492a-9a4b-4e511761edac" (UID: "f7c46fbd-c2df-492a-9a4b-4e511761edac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:51:13 crc kubenswrapper[4715]: I1203 22:51:13.015995 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7c46fbd-c2df-492a-9a4b-4e511761edac-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 22:51:13 crc kubenswrapper[4715]: I1203 22:51:13.652599 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gmrmp" Dec 03 22:51:13 crc kubenswrapper[4715]: I1203 22:51:13.689221 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gmrmp"] Dec 03 22:51:13 crc kubenswrapper[4715]: I1203 22:51:13.701230 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gmrmp"] Dec 03 22:51:15 crc kubenswrapper[4715]: I1203 22:51:15.646864 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" path="/var/lib/kubelet/pods/f7c46fbd-c2df-492a-9a4b-4e511761edac/volumes" Dec 03 22:52:35 crc kubenswrapper[4715]: I1203 22:52:35.160293 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:52:35 crc kubenswrapper[4715]: I1203 22:52:35.160903 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:53:05 crc kubenswrapper[4715]: I1203 22:53:05.159933 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:53:05 crc kubenswrapper[4715]: I1203 22:53:05.160616 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.159940 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.160450 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.160519 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.161404 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.161474 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" gracePeriod=600 Dec 03 22:53:35 crc kubenswrapper[4715]: E1203 22:53:35.480718 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.988862 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" exitCode=0 Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.988922 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828"} Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.988958 4715 scope.go:117] "RemoveContainer" containerID="f37b824df953b26abbc5aaf2b2af398da5e300d4610032a787f3edd67e7adf4e" Dec 03 22:53:35 crc kubenswrapper[4715]: I1203 22:53:35.989541 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:53:35 crc kubenswrapper[4715]: E1203 22:53:35.989770 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:53:50 crc kubenswrapper[4715]: I1203 22:53:50.633993 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:53:50 crc kubenswrapper[4715]: E1203 22:53:50.634695 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:54:02 crc kubenswrapper[4715]: I1203 22:54:02.635304 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:54:02 crc kubenswrapper[4715]: E1203 22:54:02.636428 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:54:13 crc kubenswrapper[4715]: I1203 22:54:13.641855 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:54:13 crc kubenswrapper[4715]: E1203 22:54:13.642795 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:54:26 crc kubenswrapper[4715]: I1203 22:54:26.634355 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:54:26 crc kubenswrapper[4715]: E1203 22:54:26.635050 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:54:39 crc kubenswrapper[4715]: I1203 22:54:39.634618 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:54:39 crc kubenswrapper[4715]: E1203 22:54:39.635303 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:54:52 crc kubenswrapper[4715]: I1203 22:54:52.636161 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:54:52 crc kubenswrapper[4715]: E1203 22:54:52.637130 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:55:05 crc kubenswrapper[4715]: I1203 22:55:05.635650 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:55:05 crc kubenswrapper[4715]: E1203 22:55:05.637391 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:55:17 crc kubenswrapper[4715]: I1203 22:55:17.634247 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:55:17 crc kubenswrapper[4715]: E1203 22:55:17.634985 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:55:28 crc kubenswrapper[4715]: I1203 22:55:28.634937 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:55:28 crc kubenswrapper[4715]: E1203 22:55:28.635740 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:55:41 crc kubenswrapper[4715]: I1203 22:55:41.635763 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:55:41 crc kubenswrapper[4715]: E1203 22:55:41.637276 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:55:55 crc kubenswrapper[4715]: I1203 22:55:55.634698 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:55:55 crc kubenswrapper[4715]: E1203 22:55:55.635473 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:55:56 crc kubenswrapper[4715]: I1203 22:55:56.217775 4715 scope.go:117] "RemoveContainer" containerID="d3ba6423c0961fc2e64fefc132c4efd224d3dab3f4cd51286c1bb87ac1628cdd" Dec 03 22:55:56 crc kubenswrapper[4715]: I1203 22:55:56.247975 4715 scope.go:117] "RemoveContainer" containerID="e3231676ed3fa1f6d9425316a6810e5fb432711273af96d8c6ecc5f6326d52be" Dec 03 22:55:56 crc kubenswrapper[4715]: I1203 22:55:56.296154 4715 scope.go:117] "RemoveContainer" containerID="abf068aad8ca7c6f1fcf055463eb7680ea3b5a5d5e989d5d36df749f0694ad75" Dec 03 22:56:09 crc kubenswrapper[4715]: I1203 22:56:09.635311 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:56:09 crc kubenswrapper[4715]: E1203 22:56:09.636030 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:56:23 crc kubenswrapper[4715]: I1203 22:56:23.643177 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:56:23 crc kubenswrapper[4715]: E1203 22:56:23.644013 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:56:36 crc kubenswrapper[4715]: I1203 22:56:36.634391 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:56:36 crc kubenswrapper[4715]: E1203 22:56:36.635443 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:56:49 crc kubenswrapper[4715]: I1203 22:56:49.636269 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:56:49 crc kubenswrapper[4715]: E1203 22:56:49.637692 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:56:56 crc kubenswrapper[4715]: I1203 22:56:56.347598 4715 scope.go:117] "RemoveContainer" containerID="66560681be74972d7a786fdb4a7e6464caaec5fe81c194980653d9a8d62de66f" Dec 03 22:56:56 crc kubenswrapper[4715]: I1203 22:56:56.379287 4715 scope.go:117] "RemoveContainer" containerID="fa0ecc8b2f0ddce0313119bd2bce817d8548fc218ae233aa56272fd2535494f7" Dec 03 22:56:56 crc kubenswrapper[4715]: I1203 22:56:56.440745 4715 scope.go:117] "RemoveContainer" containerID="2849fba46a997fb2750dd28b2db32e16537e7cc8cfac1f9f7e9c15342c8e898e" Dec 03 22:56:56 crc kubenswrapper[4715]: I1203 22:56:56.895696 4715 generic.go:334] "Generic (PLEG): container finished" podID="3c5223ae-7341-4aa4-821b-ec5450e37ebc" containerID="0b0e231b4437de350c57aa874ba905e6030ee222655536f4c8df0d44e621428a" exitCode=0 Dec 03 22:56:56 crc kubenswrapper[4715]: I1203 22:56:56.895743 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3c5223ae-7341-4aa4-821b-ec5450e37ebc","Type":"ContainerDied","Data":"0b0e231b4437de350c57aa874ba905e6030ee222655536f4c8df0d44e621428a"} Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.502244 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.692765 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config-secret\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.692834 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-config-data\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.692885 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.692917 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-temporary\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.692968 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ssh-key\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.693006 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ca-certs\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.693041 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-workdir\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.694074 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.694010 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.694344 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pvhs\" (UniqueName: \"kubernetes.io/projected/3c5223ae-7341-4aa4-821b-ec5450e37ebc-kube-api-access-2pvhs\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.694738 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-config-data" (OuterVolumeSpecName: "config-data") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.697306 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.697354 4715 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.699243 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.700260 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.716882 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c5223ae-7341-4aa4-821b-ec5450e37ebc-kube-api-access-2pvhs" (OuterVolumeSpecName: "kube-api-access-2pvhs") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "kube-api-access-2pvhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.728913 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.751432 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.761955 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.798617 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.799049 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config\") pod \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\" (UID: \"3c5223ae-7341-4aa4-821b-ec5450e37ebc\") " Dec 03 22:56:58 crc kubenswrapper[4715]: W1203 22:56:58.799142 4715 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/3c5223ae-7341-4aa4-821b-ec5450e37ebc/volumes/kubernetes.io~configmap/openstack-config Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.799157 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "3c5223ae-7341-4aa4-821b-ec5450e37ebc" (UID: "3c5223ae-7341-4aa4-821b-ec5450e37ebc"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.799842 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pvhs\" (UniqueName: \"kubernetes.io/projected/3c5223ae-7341-4aa4-821b-ec5450e37ebc-kube-api-access-2pvhs\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.799879 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.799895 4715 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3c5223ae-7341-4aa4-821b-ec5450e37ebc-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.799911 4715 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.800686 4715 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3c5223ae-7341-4aa4-821b-ec5450e37ebc-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.800788 4715 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3c5223ae-7341-4aa4-821b-ec5450e37ebc-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.800835 4715 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.823593 4715 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.902618 4715 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.919996 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3c5223ae-7341-4aa4-821b-ec5450e37ebc","Type":"ContainerDied","Data":"51b3247ed5534bf356fa4fb44544509b08c524ffb025382b89d0a0068757573a"} Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.920058 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51b3247ed5534bf356fa4fb44544509b08c524ffb025382b89d0a0068757573a" Dec 03 22:56:58 crc kubenswrapper[4715]: I1203 22:56:58.920058 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 22:57:03 crc kubenswrapper[4715]: I1203 22:57:03.642098 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:57:03 crc kubenswrapper[4715]: E1203 22:57:03.643022 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.208968 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 22:57:10 crc kubenswrapper[4715]: E1203 22:57:10.209965 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerName="extract-utilities" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.209982 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerName="extract-utilities" Dec 03 22:57:10 crc kubenswrapper[4715]: E1203 22:57:10.209996 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="extract-utilities" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210004 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="extract-utilities" Dec 03 22:57:10 crc kubenswrapper[4715]: E1203 22:57:10.210017 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="extract-content" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210026 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="extract-content" Dec 03 22:57:10 crc kubenswrapper[4715]: E1203 22:57:10.210045 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c5223ae-7341-4aa4-821b-ec5450e37ebc" containerName="tempest-tests-tempest-tests-runner" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210053 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c5223ae-7341-4aa4-821b-ec5450e37ebc" containerName="tempest-tests-tempest-tests-runner" Dec 03 22:57:10 crc kubenswrapper[4715]: E1203 22:57:10.210072 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerName="registry-server" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210081 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerName="registry-server" Dec 03 22:57:10 crc kubenswrapper[4715]: E1203 22:57:10.210096 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerName="extract-content" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210103 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerName="extract-content" Dec 03 22:57:10 crc kubenswrapper[4715]: E1203 22:57:10.210121 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="registry-server" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210129 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="registry-server" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210346 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="87d892b8-24ea-4e58-ac8a-2ed88e721791" containerName="registry-server" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210375 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c5223ae-7341-4aa4-821b-ec5450e37ebc" containerName="tempest-tests-tempest-tests-runner" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.210393 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7c46fbd-c2df-492a-9a4b-4e511761edac" containerName="registry-server" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.211156 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.214282 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-7f5d9" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.217433 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.220951 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"bf66462e-6db6-4370-9071-d4d60593015d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.221050 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ffc5\" (UniqueName: \"kubernetes.io/projected/bf66462e-6db6-4370-9071-d4d60593015d-kube-api-access-4ffc5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"bf66462e-6db6-4370-9071-d4d60593015d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.322588 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ffc5\" (UniqueName: \"kubernetes.io/projected/bf66462e-6db6-4370-9071-d4d60593015d-kube-api-access-4ffc5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"bf66462e-6db6-4370-9071-d4d60593015d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.322891 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"bf66462e-6db6-4370-9071-d4d60593015d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.323390 4715 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"bf66462e-6db6-4370-9071-d4d60593015d\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.344960 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ffc5\" (UniqueName: \"kubernetes.io/projected/bf66462e-6db6-4370-9071-d4d60593015d-kube-api-access-4ffc5\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"bf66462e-6db6-4370-9071-d4d60593015d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.348686 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"bf66462e-6db6-4370-9071-d4d60593015d\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.535221 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.971059 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 22:57:10 crc kubenswrapper[4715]: I1203 22:57:10.978274 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 22:57:11 crc kubenswrapper[4715]: I1203 22:57:11.030798 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"bf66462e-6db6-4370-9071-d4d60593015d","Type":"ContainerStarted","Data":"045687aa598931326c68585d9aaf296abd26facb1a802cf71bddd90e74ef715e"} Dec 03 22:57:13 crc kubenswrapper[4715]: I1203 22:57:13.050311 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"bf66462e-6db6-4370-9071-d4d60593015d","Type":"ContainerStarted","Data":"7fab1fd60c8735ad9c981912f0063144d654a17ae6fd2450ca75faae0261ebcc"} Dec 03 22:57:13 crc kubenswrapper[4715]: I1203 22:57:13.073778 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.820922106 podStartE2EDuration="3.073760573s" podCreationTimestamp="2025-12-03 22:57:10 +0000 UTC" firstStartedPulling="2025-12-03 22:57:10.977970785 +0000 UTC m=+4527.720681390" lastFinishedPulling="2025-12-03 22:57:12.230809262 +0000 UTC m=+4528.973519857" observedRunningTime="2025-12-03 22:57:13.06448689 +0000 UTC m=+4529.807197485" watchObservedRunningTime="2025-12-03 22:57:13.073760573 +0000 UTC m=+4529.816471168" Dec 03 22:57:15 crc kubenswrapper[4715]: I1203 22:57:15.634426 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:57:15 crc kubenswrapper[4715]: E1203 22:57:15.635340 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:57:29 crc kubenswrapper[4715]: I1203 22:57:29.634338 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:57:29 crc kubenswrapper[4715]: E1203 22:57:29.635144 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.431260 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kj54w/must-gather-v5xct"] Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.434554 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.438058 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-kj54w"/"default-dockercfg-rfgnl" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.438593 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kj54w"/"openshift-service-ca.crt" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.438708 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-kj54w"/"kube-root-ca.crt" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.442360 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kj54w/must-gather-v5xct"] Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.522251 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d0ca9866-708a-48c1-847b-bc7cb8000b02-must-gather-output\") pod \"must-gather-v5xct\" (UID: \"d0ca9866-708a-48c1-847b-bc7cb8000b02\") " pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.522296 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmksc\" (UniqueName: \"kubernetes.io/projected/d0ca9866-708a-48c1-847b-bc7cb8000b02-kube-api-access-wmksc\") pod \"must-gather-v5xct\" (UID: \"d0ca9866-708a-48c1-847b-bc7cb8000b02\") " pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.624645 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d0ca9866-708a-48c1-847b-bc7cb8000b02-must-gather-output\") pod \"must-gather-v5xct\" (UID: \"d0ca9866-708a-48c1-847b-bc7cb8000b02\") " pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.625015 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmksc\" (UniqueName: \"kubernetes.io/projected/d0ca9866-708a-48c1-847b-bc7cb8000b02-kube-api-access-wmksc\") pod \"must-gather-v5xct\" (UID: \"d0ca9866-708a-48c1-847b-bc7cb8000b02\") " pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.625139 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d0ca9866-708a-48c1-847b-bc7cb8000b02-must-gather-output\") pod \"must-gather-v5xct\" (UID: \"d0ca9866-708a-48c1-847b-bc7cb8000b02\") " pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.647376 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmksc\" (UniqueName: \"kubernetes.io/projected/d0ca9866-708a-48c1-847b-bc7cb8000b02-kube-api-access-wmksc\") pod \"must-gather-v5xct\" (UID: \"d0ca9866-708a-48c1-847b-bc7cb8000b02\") " pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 22:57:36 crc kubenswrapper[4715]: I1203 22:57:36.752195 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 22:57:37 crc kubenswrapper[4715]: I1203 22:57:37.234109 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-kj54w/must-gather-v5xct"] Dec 03 22:57:37 crc kubenswrapper[4715]: I1203 22:57:37.262481 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/must-gather-v5xct" event={"ID":"d0ca9866-708a-48c1-847b-bc7cb8000b02","Type":"ContainerStarted","Data":"37a9859896b88680b3fb213b672b332519016e866900f00c2a5c4bb4c45196a4"} Dec 03 22:57:40 crc kubenswrapper[4715]: I1203 22:57:40.634662 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:57:40 crc kubenswrapper[4715]: E1203 22:57:40.635279 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:57:48 crc kubenswrapper[4715]: I1203 22:57:48.351510 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/must-gather-v5xct" event={"ID":"d0ca9866-708a-48c1-847b-bc7cb8000b02","Type":"ContainerStarted","Data":"e6f96deada2c7c60fab4e6931faf7060d7b80dac52e23bdf2286820864a70e8d"} Dec 03 22:57:48 crc kubenswrapper[4715]: I1203 22:57:48.351949 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/must-gather-v5xct" event={"ID":"d0ca9866-708a-48c1-847b-bc7cb8000b02","Type":"ContainerStarted","Data":"4ecbfed57d7dd4008ef3432170d724b5f180638eb6db310a58d80b788dd0d852"} Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.392457 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-kj54w/must-gather-v5xct" podStartSLOduration=5.167643466 podStartE2EDuration="15.392435919s" podCreationTimestamp="2025-12-03 22:57:36 +0000 UTC" firstStartedPulling="2025-12-03 22:57:37.235903871 +0000 UTC m=+4553.978614466" lastFinishedPulling="2025-12-03 22:57:47.460696324 +0000 UTC m=+4564.203406919" observedRunningTime="2025-12-03 22:57:48.37325622 +0000 UTC m=+4565.115966815" watchObservedRunningTime="2025-12-03 22:57:51.392435919 +0000 UTC m=+4568.135146514" Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.404184 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kj54w/crc-debug-rqvp2"] Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.405719 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.537070 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrn5w\" (UniqueName: \"kubernetes.io/projected/ad752f37-de33-4809-8308-50c084b7ed20-kube-api-access-wrn5w\") pod \"crc-debug-rqvp2\" (UID: \"ad752f37-de33-4809-8308-50c084b7ed20\") " pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.537166 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad752f37-de33-4809-8308-50c084b7ed20-host\") pod \"crc-debug-rqvp2\" (UID: \"ad752f37-de33-4809-8308-50c084b7ed20\") " pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.638842 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrn5w\" (UniqueName: \"kubernetes.io/projected/ad752f37-de33-4809-8308-50c084b7ed20-kube-api-access-wrn5w\") pod \"crc-debug-rqvp2\" (UID: \"ad752f37-de33-4809-8308-50c084b7ed20\") " pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.638933 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad752f37-de33-4809-8308-50c084b7ed20-host\") pod \"crc-debug-rqvp2\" (UID: \"ad752f37-de33-4809-8308-50c084b7ed20\") " pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.639019 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad752f37-de33-4809-8308-50c084b7ed20-host\") pod \"crc-debug-rqvp2\" (UID: \"ad752f37-de33-4809-8308-50c084b7ed20\") " pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.657479 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrn5w\" (UniqueName: \"kubernetes.io/projected/ad752f37-de33-4809-8308-50c084b7ed20-kube-api-access-wrn5w\") pod \"crc-debug-rqvp2\" (UID: \"ad752f37-de33-4809-8308-50c084b7ed20\") " pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:57:51 crc kubenswrapper[4715]: I1203 22:57:51.726159 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:57:51 crc kubenswrapper[4715]: W1203 22:57:51.753361 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad752f37_de33_4809_8308_50c084b7ed20.slice/crio-5512f8f2dacdc4d5c416a01f9a06cca77306768a096a4ba75286d45b7dd5e634 WatchSource:0}: Error finding container 5512f8f2dacdc4d5c416a01f9a06cca77306768a096a4ba75286d45b7dd5e634: Status 404 returned error can't find the container with id 5512f8f2dacdc4d5c416a01f9a06cca77306768a096a4ba75286d45b7dd5e634 Dec 03 22:57:52 crc kubenswrapper[4715]: I1203 22:57:52.381119 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/crc-debug-rqvp2" event={"ID":"ad752f37-de33-4809-8308-50c084b7ed20","Type":"ContainerStarted","Data":"5512f8f2dacdc4d5c416a01f9a06cca77306768a096a4ba75286d45b7dd5e634"} Dec 03 22:57:55 crc kubenswrapper[4715]: I1203 22:57:55.635178 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:57:55 crc kubenswrapper[4715]: E1203 22:57:55.635896 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:58:04 crc kubenswrapper[4715]: I1203 22:58:04.486148 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/crc-debug-rqvp2" event={"ID":"ad752f37-de33-4809-8308-50c084b7ed20","Type":"ContainerStarted","Data":"88b53ac05184d80eb238e26e17b9d0d6587ca98597bb83ae276dde0c50e9fc90"} Dec 03 22:58:04 crc kubenswrapper[4715]: I1203 22:58:04.518444 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-kj54w/crc-debug-rqvp2" podStartSLOduration=1.073733061 podStartE2EDuration="13.518416611s" podCreationTimestamp="2025-12-03 22:57:51 +0000 UTC" firstStartedPulling="2025-12-03 22:57:51.755838249 +0000 UTC m=+4568.498548844" lastFinishedPulling="2025-12-03 22:58:04.200521799 +0000 UTC m=+4580.943232394" observedRunningTime="2025-12-03 22:58:04.513087115 +0000 UTC m=+4581.255797710" watchObservedRunningTime="2025-12-03 22:58:04.518416611 +0000 UTC m=+4581.261127206" Dec 03 22:58:08 crc kubenswrapper[4715]: I1203 22:58:08.635252 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:58:08 crc kubenswrapper[4715]: E1203 22:58:08.636013 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:58:23 crc kubenswrapper[4715]: I1203 22:58:23.644136 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:58:23 crc kubenswrapper[4715]: E1203 22:58:23.645010 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:58:34 crc kubenswrapper[4715]: I1203 22:58:34.634672 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:58:34 crc kubenswrapper[4715]: E1203 22:58:34.635482 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 22:58:45 crc kubenswrapper[4715]: I1203 22:58:45.634484 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 22:58:45 crc kubenswrapper[4715]: I1203 22:58:45.881886 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"0ba545ca47db831472f2241785c844e2b365936c2d043769501a6e2a2e23232d"} Dec 03 22:58:48 crc kubenswrapper[4715]: I1203 22:58:48.909636 4715 generic.go:334] "Generic (PLEG): container finished" podID="ad752f37-de33-4809-8308-50c084b7ed20" containerID="88b53ac05184d80eb238e26e17b9d0d6587ca98597bb83ae276dde0c50e9fc90" exitCode=0 Dec 03 22:58:48 crc kubenswrapper[4715]: I1203 22:58:48.909742 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/crc-debug-rqvp2" event={"ID":"ad752f37-de33-4809-8308-50c084b7ed20","Type":"ContainerDied","Data":"88b53ac05184d80eb238e26e17b9d0d6587ca98597bb83ae276dde0c50e9fc90"} Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.032015 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.066598 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kj54w/crc-debug-rqvp2"] Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.073970 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kj54w/crc-debug-rqvp2"] Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.123424 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrn5w\" (UniqueName: \"kubernetes.io/projected/ad752f37-de33-4809-8308-50c084b7ed20-kube-api-access-wrn5w\") pod \"ad752f37-de33-4809-8308-50c084b7ed20\" (UID: \"ad752f37-de33-4809-8308-50c084b7ed20\") " Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.123640 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad752f37-de33-4809-8308-50c084b7ed20-host\") pod \"ad752f37-de33-4809-8308-50c084b7ed20\" (UID: \"ad752f37-de33-4809-8308-50c084b7ed20\") " Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.124278 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ad752f37-de33-4809-8308-50c084b7ed20-host" (OuterVolumeSpecName: "host") pod "ad752f37-de33-4809-8308-50c084b7ed20" (UID: "ad752f37-de33-4809-8308-50c084b7ed20"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.132386 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad752f37-de33-4809-8308-50c084b7ed20-kube-api-access-wrn5w" (OuterVolumeSpecName: "kube-api-access-wrn5w") pod "ad752f37-de33-4809-8308-50c084b7ed20" (UID: "ad752f37-de33-4809-8308-50c084b7ed20"). InnerVolumeSpecName "kube-api-access-wrn5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.226144 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrn5w\" (UniqueName: \"kubernetes.io/projected/ad752f37-de33-4809-8308-50c084b7ed20-kube-api-access-wrn5w\") on node \"crc\" DevicePath \"\"" Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.226185 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ad752f37-de33-4809-8308-50c084b7ed20-host\") on node \"crc\" DevicePath \"\"" Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.931601 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5512f8f2dacdc4d5c416a01f9a06cca77306768a096a4ba75286d45b7dd5e634" Dec 03 22:58:50 crc kubenswrapper[4715]: I1203 22:58:50.931669 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-rqvp2" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.217968 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kj54w/crc-debug-ps5dl"] Dec 03 22:58:51 crc kubenswrapper[4715]: E1203 22:58:51.218641 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad752f37-de33-4809-8308-50c084b7ed20" containerName="container-00" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.218653 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad752f37-de33-4809-8308-50c084b7ed20" containerName="container-00" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.218846 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad752f37-de33-4809-8308-50c084b7ed20" containerName="container-00" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.219474 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.347084 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-host\") pod \"crc-debug-ps5dl\" (UID: \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\") " pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.347242 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tck8s\" (UniqueName: \"kubernetes.io/projected/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-kube-api-access-tck8s\") pod \"crc-debug-ps5dl\" (UID: \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\") " pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.449146 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-host\") pod \"crc-debug-ps5dl\" (UID: \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\") " pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.449429 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tck8s\" (UniqueName: \"kubernetes.io/projected/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-kube-api-access-tck8s\") pod \"crc-debug-ps5dl\" (UID: \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\") " pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.449765 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-host\") pod \"crc-debug-ps5dl\" (UID: \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\") " pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.467839 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tck8s\" (UniqueName: \"kubernetes.io/projected/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-kube-api-access-tck8s\") pod \"crc-debug-ps5dl\" (UID: \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\") " pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.541558 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.648381 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad752f37-de33-4809-8308-50c084b7ed20" path="/var/lib/kubelet/pods/ad752f37-de33-4809-8308-50c084b7ed20/volumes" Dec 03 22:58:51 crc kubenswrapper[4715]: I1203 22:58:51.940337 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/crc-debug-ps5dl" event={"ID":"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a","Type":"ContainerStarted","Data":"1643ff864605372498bced50a9ed0785cf1727948a527985a9383cba30126b92"} Dec 03 22:58:53 crc kubenswrapper[4715]: I1203 22:58:53.960674 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/crc-debug-ps5dl" event={"ID":"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a","Type":"ContainerStarted","Data":"9b89e649cd2d6490e6640ddf66447325c6741196bf6f2556bcb3709401a34413"} Dec 03 22:58:54 crc kubenswrapper[4715]: I1203 22:58:54.972149 4715 generic.go:334] "Generic (PLEG): container finished" podID="6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a" containerID="9b89e649cd2d6490e6640ddf66447325c6741196bf6f2556bcb3709401a34413" exitCode=0 Dec 03 22:58:54 crc kubenswrapper[4715]: I1203 22:58:54.972209 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/crc-debug-ps5dl" event={"ID":"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a","Type":"ContainerDied","Data":"9b89e649cd2d6490e6640ddf66447325c6741196bf6f2556bcb3709401a34413"} Dec 03 22:58:55 crc kubenswrapper[4715]: I1203 22:58:55.470240 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kj54w/crc-debug-ps5dl"] Dec 03 22:58:55 crc kubenswrapper[4715]: I1203 22:58:55.479470 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kj54w/crc-debug-ps5dl"] Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.076495 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.242975 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-host\") pod \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\" (UID: \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\") " Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.243059 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-host" (OuterVolumeSpecName: "host") pod "6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a" (UID: "6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.243411 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tck8s\" (UniqueName: \"kubernetes.io/projected/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-kube-api-access-tck8s\") pod \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\" (UID: \"6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a\") " Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.243883 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-host\") on node \"crc\" DevicePath \"\"" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.249161 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-kube-api-access-tck8s" (OuterVolumeSpecName: "kube-api-access-tck8s") pod "6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a" (UID: "6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a"). InnerVolumeSpecName "kube-api-access-tck8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.345304 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tck8s\" (UniqueName: \"kubernetes.io/projected/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a-kube-api-access-tck8s\") on node \"crc\" DevicePath \"\"" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.661153 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-kj54w/crc-debug-89h5f"] Dec 03 22:58:56 crc kubenswrapper[4715]: E1203 22:58:56.661719 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a" containerName="container-00" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.661753 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a" containerName="container-00" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.662082 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a" containerName="container-00" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.662900 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.752483 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49q7p\" (UniqueName: \"kubernetes.io/projected/2d0af008-8bbc-457d-8225-28217d6cb3f8-kube-api-access-49q7p\") pod \"crc-debug-89h5f\" (UID: \"2d0af008-8bbc-457d-8225-28217d6cb3f8\") " pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.752590 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d0af008-8bbc-457d-8225-28217d6cb3f8-host\") pod \"crc-debug-89h5f\" (UID: \"2d0af008-8bbc-457d-8225-28217d6cb3f8\") " pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.855021 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49q7p\" (UniqueName: \"kubernetes.io/projected/2d0af008-8bbc-457d-8225-28217d6cb3f8-kube-api-access-49q7p\") pod \"crc-debug-89h5f\" (UID: \"2d0af008-8bbc-457d-8225-28217d6cb3f8\") " pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.855416 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d0af008-8bbc-457d-8225-28217d6cb3f8-host\") pod \"crc-debug-89h5f\" (UID: \"2d0af008-8bbc-457d-8225-28217d6cb3f8\") " pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.855552 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d0af008-8bbc-457d-8225-28217d6cb3f8-host\") pod \"crc-debug-89h5f\" (UID: \"2d0af008-8bbc-457d-8225-28217d6cb3f8\") " pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.871868 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49q7p\" (UniqueName: \"kubernetes.io/projected/2d0af008-8bbc-457d-8225-28217d6cb3f8-kube-api-access-49q7p\") pod \"crc-debug-89h5f\" (UID: \"2d0af008-8bbc-457d-8225-28217d6cb3f8\") " pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.981106 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.992457 4715 scope.go:117] "RemoveContainer" containerID="9b89e649cd2d6490e6640ddf66447325c6741196bf6f2556bcb3709401a34413" Dec 03 22:58:56 crc kubenswrapper[4715]: I1203 22:58:56.992512 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-ps5dl" Dec 03 22:58:57 crc kubenswrapper[4715]: I1203 22:58:57.645920 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a" path="/var/lib/kubelet/pods/6f7c90e1-5d8d-4c67-85dc-8a3b3b8f733a/volumes" Dec 03 22:58:58 crc kubenswrapper[4715]: I1203 22:58:58.003531 4715 generic.go:334] "Generic (PLEG): container finished" podID="2d0af008-8bbc-457d-8225-28217d6cb3f8" containerID="0db5b6133db18a5e882e3e4cb66a455fc9f28c7fcf65f81660a87bc11a240135" exitCode=0 Dec 03 22:58:58 crc kubenswrapper[4715]: I1203 22:58:58.003572 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/crc-debug-89h5f" event={"ID":"2d0af008-8bbc-457d-8225-28217d6cb3f8","Type":"ContainerDied","Data":"0db5b6133db18a5e882e3e4cb66a455fc9f28c7fcf65f81660a87bc11a240135"} Dec 03 22:58:58 crc kubenswrapper[4715]: I1203 22:58:58.003598 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/crc-debug-89h5f" event={"ID":"2d0af008-8bbc-457d-8225-28217d6cb3f8","Type":"ContainerStarted","Data":"3ff9a65f6a5c01a1c4bf1cab1fcd952e77cdbb4bdf45268abbed9aecaaf03231"} Dec 03 22:58:58 crc kubenswrapper[4715]: I1203 22:58:58.046081 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kj54w/crc-debug-89h5f"] Dec 03 22:58:58 crc kubenswrapper[4715]: I1203 22:58:58.054141 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kj54w/crc-debug-89h5f"] Dec 03 22:58:59 crc kubenswrapper[4715]: I1203 22:58:59.112663 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:58:59 crc kubenswrapper[4715]: I1203 22:58:59.202559 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49q7p\" (UniqueName: \"kubernetes.io/projected/2d0af008-8bbc-457d-8225-28217d6cb3f8-kube-api-access-49q7p\") pod \"2d0af008-8bbc-457d-8225-28217d6cb3f8\" (UID: \"2d0af008-8bbc-457d-8225-28217d6cb3f8\") " Dec 03 22:58:59 crc kubenswrapper[4715]: I1203 22:58:59.202744 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d0af008-8bbc-457d-8225-28217d6cb3f8-host\") pod \"2d0af008-8bbc-457d-8225-28217d6cb3f8\" (UID: \"2d0af008-8bbc-457d-8225-28217d6cb3f8\") " Dec 03 22:58:59 crc kubenswrapper[4715]: I1203 22:58:59.202849 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d0af008-8bbc-457d-8225-28217d6cb3f8-host" (OuterVolumeSpecName: "host") pod "2d0af008-8bbc-457d-8225-28217d6cb3f8" (UID: "2d0af008-8bbc-457d-8225-28217d6cb3f8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 22:58:59 crc kubenswrapper[4715]: I1203 22:58:59.203291 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d0af008-8bbc-457d-8225-28217d6cb3f8-host\") on node \"crc\" DevicePath \"\"" Dec 03 22:58:59 crc kubenswrapper[4715]: I1203 22:58:59.214757 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d0af008-8bbc-457d-8225-28217d6cb3f8-kube-api-access-49q7p" (OuterVolumeSpecName: "kube-api-access-49q7p") pod "2d0af008-8bbc-457d-8225-28217d6cb3f8" (UID: "2d0af008-8bbc-457d-8225-28217d6cb3f8"). InnerVolumeSpecName "kube-api-access-49q7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 22:58:59 crc kubenswrapper[4715]: I1203 22:58:59.304724 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49q7p\" (UniqueName: \"kubernetes.io/projected/2d0af008-8bbc-457d-8225-28217d6cb3f8-kube-api-access-49q7p\") on node \"crc\" DevicePath \"\"" Dec 03 22:58:59 crc kubenswrapper[4715]: I1203 22:58:59.645088 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d0af008-8bbc-457d-8225-28217d6cb3f8" path="/var/lib/kubelet/pods/2d0af008-8bbc-457d-8225-28217d6cb3f8/volumes" Dec 03 22:59:00 crc kubenswrapper[4715]: I1203 22:59:00.025528 4715 scope.go:117] "RemoveContainer" containerID="0db5b6133db18a5e882e3e4cb66a455fc9f28c7fcf65f81660a87bc11a240135" Dec 03 22:59:00 crc kubenswrapper[4715]: I1203 22:59:00.025624 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/crc-debug-89h5f" Dec 03 22:59:13 crc kubenswrapper[4715]: I1203 22:59:13.509767 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-dd5bb56d6-jb5n2_9170a301-98ac-4953-ae87-4910bf0a466a/barbican-api/0.log" Dec 03 22:59:13 crc kubenswrapper[4715]: I1203 22:59:13.655755 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-dd5bb56d6-jb5n2_9170a301-98ac-4953-ae87-4910bf0a466a/barbican-api-log/0.log" Dec 03 22:59:13 crc kubenswrapper[4715]: I1203 22:59:13.720732 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9cc86b6b-6xlfj_0e740246-a3e9-4692-8888-e12768f0573e/barbican-keystone-listener/0.log" Dec 03 22:59:13 crc kubenswrapper[4715]: I1203 22:59:13.738438 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9cc86b6b-6xlfj_0e740246-a3e9-4692-8888-e12768f0573e/barbican-keystone-listener-log/0.log" Dec 03 22:59:13 crc kubenswrapper[4715]: I1203 22:59:13.954106 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-78449cf9c5-v4vnx_4382c9dd-8a2d-4634-b9de-0796a994d491/barbican-worker/0.log" Dec 03 22:59:13 crc kubenswrapper[4715]: I1203 22:59:13.981385 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-78449cf9c5-v4vnx_4382c9dd-8a2d-4634-b9de-0796a994d491/barbican-worker-log/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.208792 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx_7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.232138 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ac3428ad-2684-4ca1-ab1e-254306dc84a3/ceilometer-central-agent/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.362457 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ac3428ad-2684-4ca1-ab1e-254306dc84a3/ceilometer-notification-agent/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.407992 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ac3428ad-2684-4ca1-ab1e-254306dc84a3/proxy-httpd/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.459525 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ac3428ad-2684-4ca1-ab1e-254306dc84a3/sg-core/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.600092 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6cfc1c9d-e789-449f-a73d-a4aead3a03f1/cinder-api/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.638490 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6cfc1c9d-e789-449f-a73d-a4aead3a03f1/cinder-api-log/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.860721 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cad3e724-7c3e-498d-9a7a-362a4f57c2b5/probe/0.log" Dec 03 22:59:14 crc kubenswrapper[4715]: I1203 22:59:14.881142 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cad3e724-7c3e-498d-9a7a-362a4f57c2b5/cinder-scheduler/0.log" Dec 03 22:59:15 crc kubenswrapper[4715]: I1203 22:59:15.069863 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg_7fa5989a-f985-4757-84dd-3c658472a85e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:15 crc kubenswrapper[4715]: I1203 22:59:15.133677 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-9mszz_b1a79612-747c-4bbd-be8c-f79a9da177db/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:15 crc kubenswrapper[4715]: I1203 22:59:15.595200 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-bhvj6_a1e5e31a-ff62-48c9-8908-15ff570bda0d/init/0.log" Dec 03 22:59:15 crc kubenswrapper[4715]: I1203 22:59:15.752870 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-bhvj6_a1e5e31a-ff62-48c9-8908-15ff570bda0d/init/0.log" Dec 03 22:59:15 crc kubenswrapper[4715]: I1203 22:59:15.806153 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-bhvj6_a1e5e31a-ff62-48c9-8908-15ff570bda0d/dnsmasq-dns/0.log" Dec 03 22:59:15 crc kubenswrapper[4715]: I1203 22:59:15.854573 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5_54307606-6028-435b-bb67-3dd55f7a938a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:16 crc kubenswrapper[4715]: I1203 22:59:16.043379 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_99136b0c-47a6-4503-806e-ec70fa8c2409/glance-httpd/0.log" Dec 03 22:59:16 crc kubenswrapper[4715]: I1203 22:59:16.101138 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_99136b0c-47a6-4503-806e-ec70fa8c2409/glance-log/0.log" Dec 03 22:59:16 crc kubenswrapper[4715]: I1203 22:59:16.275257 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_20ba12ff-26f9-4100-b21d-0545d33fc7d1/glance-log/0.log" Dec 03 22:59:16 crc kubenswrapper[4715]: I1203 22:59:16.378561 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_20ba12ff-26f9-4100-b21d-0545d33fc7d1/glance-httpd/0.log" Dec 03 22:59:16 crc kubenswrapper[4715]: I1203 22:59:16.474311 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-656478594d-5tbrm_6c7cfacf-a283-45d3-982d-b28e62a9146b/horizon/0.log" Dec 03 22:59:16 crc kubenswrapper[4715]: I1203 22:59:16.709948 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-c27sm_c58da0a9-c5be-4018-a322-7ba8ef960d52/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:16 crc kubenswrapper[4715]: I1203 22:59:16.867792 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-656478594d-5tbrm_6c7cfacf-a283-45d3-982d-b28e62a9146b/horizon-log/0.log" Dec 03 22:59:16 crc kubenswrapper[4715]: I1203 22:59:16.942417 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-pmd4j_3976028c-4d85-4518-9151-0e7c94638d58/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:17 crc kubenswrapper[4715]: I1203 22:59:17.253180 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_11b5a8ae-b831-453c-964b-1631e5d1280c/kube-state-metrics/0.log" Dec 03 22:59:17 crc kubenswrapper[4715]: I1203 22:59:17.260159 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-76bcbf7d4d-l7qmc_954b69af-5e1b-4eec-a783-1cbe28e7c0ca/keystone-api/0.log" Dec 03 22:59:17 crc kubenswrapper[4715]: I1203 22:59:17.415529 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4_e419d6c2-a6b7-474d-aea8-6f90eb7ce85a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:17 crc kubenswrapper[4715]: I1203 22:59:17.771638 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-666d579d9-hnzld_58badebd-4a34-4430-96ee-706deee12007/neutron-api/0.log" Dec 03 22:59:17 crc kubenswrapper[4715]: I1203 22:59:17.879449 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-666d579d9-hnzld_58badebd-4a34-4430-96ee-706deee12007/neutron-httpd/0.log" Dec 03 22:59:17 crc kubenswrapper[4715]: I1203 22:59:17.992694 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d_c3439842-caec-430b-a91e-0b03443eff15/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:18 crc kubenswrapper[4715]: I1203 22:59:18.432622 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b079ad3a-bcab-48cc-99da-f7f656195558/nova-api-log/0.log" Dec 03 22:59:18 crc kubenswrapper[4715]: I1203 22:59:18.482712 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_c328932f-799e-4f62-88c7-34c17ab3c0bc/nova-cell0-conductor-conductor/0.log" Dec 03 22:59:18 crc kubenswrapper[4715]: I1203 22:59:18.717273 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b079ad3a-bcab-48cc-99da-f7f656195558/nova-api-api/0.log" Dec 03 22:59:18 crc kubenswrapper[4715]: I1203 22:59:18.858239 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_5b686fe6-ab19-4e75-9328-7ae61b8b609f/nova-cell1-conductor-conductor/0.log" Dec 03 22:59:19 crc kubenswrapper[4715]: I1203 22:59:19.009721 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e4f46a99-13e6-4175-939a-3ffb54503ef2/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 22:59:19 crc kubenswrapper[4715]: I1203 22:59:19.311686 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-njnmd_10aa3b67-88a4-4a5d-9d81-096c5b944fb7/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:19 crc kubenswrapper[4715]: I1203 22:59:19.466053 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0a252fdb-1b30-4278-8bfc-471881a27f21/nova-metadata-log/0.log" Dec 03 22:59:19 crc kubenswrapper[4715]: I1203 22:59:19.782307 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b487c9d6-b74d-4db6-8ef4-caee02c83106/mysql-bootstrap/0.log" Dec 03 22:59:19 crc kubenswrapper[4715]: I1203 22:59:19.798349 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_5e79bc96-17dc-43d2-b049-da6de097be5e/nova-scheduler-scheduler/0.log" Dec 03 22:59:19 crc kubenswrapper[4715]: I1203 22:59:19.978585 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b487c9d6-b74d-4db6-8ef4-caee02c83106/mysql-bootstrap/0.log" Dec 03 22:59:20 crc kubenswrapper[4715]: I1203 22:59:20.003482 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b487c9d6-b74d-4db6-8ef4-caee02c83106/galera/0.log" Dec 03 22:59:20 crc kubenswrapper[4715]: I1203 22:59:20.722729 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0a252fdb-1b30-4278-8bfc-471881a27f21/nova-metadata-metadata/0.log" Dec 03 22:59:20 crc kubenswrapper[4715]: I1203 22:59:20.816944 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b024438-6200-45dd-9297-ae2b071dae15/mysql-bootstrap/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.038099 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b024438-6200-45dd-9297-ae2b071dae15/mysql-bootstrap/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.049467 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_205269ef-3624-4663-957b-907cf1a7cadc/openstackclient/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.051401 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b024438-6200-45dd-9297-ae2b071dae15/galera/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.310896 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5fzvj_893e6b97-27e2-4372-8fc6-90562f2781bb/ovn-controller/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.464747 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-sgsmm_473308fd-dd98-47fd-8a5e-32ab39bd6730/openstack-network-exporter/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.658277 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xq7sg_d495583b-de78-4dd5-aa7b-70e15ee3cf0e/ovsdb-server-init/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.854555 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xq7sg_d495583b-de78-4dd5-aa7b-70e15ee3cf0e/ovsdb-server-init/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.894610 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xq7sg_d495583b-de78-4dd5-aa7b-70e15ee3cf0e/ovs-vswitchd/0.log" Dec 03 22:59:21 crc kubenswrapper[4715]: I1203 22:59:21.919773 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xq7sg_d495583b-de78-4dd5-aa7b-70e15ee3cf0e/ovsdb-server/0.log" Dec 03 22:59:22 crc kubenswrapper[4715]: I1203 22:59:22.116700 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-68qjb_d0c3bcdb-424a-4da7-9917-7a7853067044/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:22 crc kubenswrapper[4715]: I1203 22:59:22.186083 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_18d99e58-0606-46f0-91ef-553c75d44ba4/openstack-network-exporter/0.log" Dec 03 22:59:22 crc kubenswrapper[4715]: I1203 22:59:22.241030 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_18d99e58-0606-46f0-91ef-553c75d44ba4/ovn-northd/0.log" Dec 03 22:59:22 crc kubenswrapper[4715]: I1203 22:59:22.390334 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_af93de35-2858-4a66-80a3-b7be3a28ff3f/ovsdbserver-nb/0.log" Dec 03 22:59:22 crc kubenswrapper[4715]: I1203 22:59:22.450310 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_af93de35-2858-4a66-80a3-b7be3a28ff3f/openstack-network-exporter/0.log" Dec 03 22:59:22 crc kubenswrapper[4715]: I1203 22:59:22.588388 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_695ea890-c971-4f5a-84ca-fbd83fa5d74a/openstack-network-exporter/0.log" Dec 03 22:59:22 crc kubenswrapper[4715]: I1203 22:59:22.651312 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_695ea890-c971-4f5a-84ca-fbd83fa5d74a/ovsdbserver-sb/0.log" Dec 03 22:59:22 crc kubenswrapper[4715]: I1203 22:59:22.806271 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8694d8458b-2x672_661b1086-fc52-475a-a449-c116baa618a2/placement-api/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.093485 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cb376763-8ff4-48ce-a374-ed1119c961e9/setup-container/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.130151 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8694d8458b-2x672_661b1086-fc52-475a-a449-c116baa618a2/placement-log/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.385804 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cb376763-8ff4-48ce-a374-ed1119c961e9/setup-container/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.420329 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_82483123-49b2-496c-96f4-2d9423e2dfeb/setup-container/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.426341 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cb376763-8ff4-48ce-a374-ed1119c961e9/rabbitmq/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.627599 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_82483123-49b2-496c-96f4-2d9423e2dfeb/setup-container/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.689877 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_82483123-49b2-496c-96f4-2d9423e2dfeb/rabbitmq/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.712413 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74_7790e81c-da01-4013-81ad-32bd29bba24f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:23 crc kubenswrapper[4715]: I1203 22:59:23.954149 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-8d7ms_e1c0be7c-5e1c-43dd-9148-30cd6bc42b93/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.046641 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9_4a15fc43-6ab8-4502-a2d3-9c837a74cb90/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.234626 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9hf4c_849e6369-17eb-4b31-a5cc-f44dc07c6e1a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.274153 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2jmsp_2e622ea8-2426-4e53-8775-3880cce092b2/ssh-known-hosts-edpm-deployment/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.482147 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5df8bbbf89-2tsr9_1bf3ff4d-5720-4151-ad53-3683f08c21c5/proxy-server/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.575692 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5df8bbbf89-2tsr9_1bf3ff4d-5720-4151-ad53-3683f08c21c5/proxy-httpd/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.648211 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-7f2gr_18eb44f1-da35-4a10-b826-03c94f6b525e/swift-ring-rebalance/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.817659 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/account-auditor/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.825819 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/account-reaper/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.869317 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/account-replicator/0.log" Dec 03 22:59:24 crc kubenswrapper[4715]: I1203 22:59:24.955310 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/account-server/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.044074 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/container-auditor/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.073283 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/container-replicator/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.179760 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/container-server/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.189317 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/container-updater/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.233127 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-auditor/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.321548 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-expirer/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.397331 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-replicator/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.416405 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-server/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.441814 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-updater/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.616043 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/rsync/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.702745 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/swift-recon-cron/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.779264 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x_32684f56-1e44-4d82-b8d2-f1193ba615d2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:25 crc kubenswrapper[4715]: I1203 22:59:25.921613 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_3c5223ae-7341-4aa4-821b-ec5450e37ebc/tempest-tests-tempest-tests-runner/0.log" Dec 03 22:59:26 crc kubenswrapper[4715]: I1203 22:59:26.001196 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_bf66462e-6db6-4370-9071-d4d60593015d/test-operator-logs-container/0.log" Dec 03 22:59:26 crc kubenswrapper[4715]: I1203 22:59:26.235990 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2_db7da37a-b417-433c-86e9-a18446b4af27/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 22:59:35 crc kubenswrapper[4715]: I1203 22:59:35.564023 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5ffb2b06-021a-49c0-abfe-6bb5c8acba3d/memcached/0.log" Dec 03 22:59:49 crc kubenswrapper[4715]: I1203 22:59:49.899826 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zd62w"] Dec 03 22:59:49 crc kubenswrapper[4715]: E1203 22:59:49.900849 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d0af008-8bbc-457d-8225-28217d6cb3f8" containerName="container-00" Dec 03 22:59:49 crc kubenswrapper[4715]: I1203 22:59:49.900867 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d0af008-8bbc-457d-8225-28217d6cb3f8" containerName="container-00" Dec 03 22:59:49 crc kubenswrapper[4715]: I1203 22:59:49.901137 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d0af008-8bbc-457d-8225-28217d6cb3f8" containerName="container-00" Dec 03 22:59:49 crc kubenswrapper[4715]: I1203 22:59:49.902541 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:49 crc kubenswrapper[4715]: I1203 22:59:49.914080 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zd62w"] Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.066164 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-catalog-content\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.066258 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p494r\" (UniqueName: \"kubernetes.io/projected/561f5f6c-67b2-4d46-b595-339ad5137f48-kube-api-access-p494r\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.066298 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-utilities\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.168173 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p494r\" (UniqueName: \"kubernetes.io/projected/561f5f6c-67b2-4d46-b595-339ad5137f48-kube-api-access-p494r\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.168243 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-utilities\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.168364 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-catalog-content\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.168836 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-utilities\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.168897 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-catalog-content\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.190254 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p494r\" (UniqueName: \"kubernetes.io/projected/561f5f6c-67b2-4d46-b595-339ad5137f48-kube-api-access-p494r\") pod \"certified-operators-zd62w\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.223271 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zd62w" Dec 03 22:59:50 crc kubenswrapper[4715]: I1203 22:59:50.714086 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zd62w"] Dec 03 22:59:51 crc kubenswrapper[4715]: I1203 22:59:51.518140 4715 generic.go:334] "Generic (PLEG): container finished" podID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerID="077df71e4964695572bfa2c51daf38b5907c35d62082bb08eb6d3081a352ed2b" exitCode=0 Dec 03 22:59:51 crc kubenswrapper[4715]: I1203 22:59:51.519256 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd62w" event={"ID":"561f5f6c-67b2-4d46-b595-339ad5137f48","Type":"ContainerDied","Data":"077df71e4964695572bfa2c51daf38b5907c35d62082bb08eb6d3081a352ed2b"} Dec 03 22:59:51 crc kubenswrapper[4715]: I1203 22:59:51.519369 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd62w" event={"ID":"561f5f6c-67b2-4d46-b595-339ad5137f48","Type":"ContainerStarted","Data":"7e348136f6c4b8f279cbff0b378a696a96440574fa4b84cf910d9a6e851dd47e"} Dec 03 22:59:52 crc kubenswrapper[4715]: I1203 22:59:52.529296 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd62w" event={"ID":"561f5f6c-67b2-4d46-b595-339ad5137f48","Type":"ContainerStarted","Data":"e931501bb7009123cd5935f42780627813fae05c12a463442102e8679aefc734"} Dec 03 22:59:53 crc kubenswrapper[4715]: I1203 22:59:53.551066 4715 generic.go:334] "Generic (PLEG): container finished" podID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerID="e931501bb7009123cd5935f42780627813fae05c12a463442102e8679aefc734" exitCode=0 Dec 03 22:59:53 crc kubenswrapper[4715]: I1203 22:59:53.551113 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd62w" event={"ID":"561f5f6c-67b2-4d46-b595-339ad5137f48","Type":"ContainerDied","Data":"e931501bb7009123cd5935f42780627813fae05c12a463442102e8679aefc734"} Dec 03 22:59:54 crc kubenswrapper[4715]: I1203 22:59:54.264345 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/util/0.log" Dec 03 22:59:54 crc kubenswrapper[4715]: I1203 22:59:54.264481 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/util/0.log" Dec 03 22:59:54 crc kubenswrapper[4715]: I1203 22:59:54.265066 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/pull/0.log" Dec 03 22:59:54 crc kubenswrapper[4715]: I1203 22:59:54.450083 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/pull/0.log" Dec 03 22:59:54 crc kubenswrapper[4715]: I1203 22:59:54.828808 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/extract/0.log" Dec 03 22:59:54 crc kubenswrapper[4715]: I1203 22:59:54.930018 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/pull/0.log" Dec 03 22:59:54 crc kubenswrapper[4715]: I1203 22:59:54.976248 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/util/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.018364 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5hzjb_46e49f12-51e7-45ca-a0cb-c7f8a40adf12/kube-rbac-proxy/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.180525 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5hzjb_46e49f12-51e7-45ca-a0cb-c7f8a40adf12/manager/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.377327 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-mnd86_9c8c4875-3433-4885-9afa-acbfea9d546b/kube-rbac-proxy/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.412359 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-mnd86_9c8c4875-3433-4885-9afa-acbfea9d546b/manager/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.500178 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-sctl9_a91ec3b7-705a-4500-be8e-e8c789a2c973/kube-rbac-proxy/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.570895 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd62w" event={"ID":"561f5f6c-67b2-4d46-b595-339ad5137f48","Type":"ContainerStarted","Data":"c1642f36dd288afa37f27a2367b9230b08ce5b1e332c93ad4648b57e15bb7971"} Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.596697 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zd62w" podStartSLOduration=2.957593984 podStartE2EDuration="6.596676533s" podCreationTimestamp="2025-12-03 22:59:49 +0000 UTC" firstStartedPulling="2025-12-03 22:59:51.521652268 +0000 UTC m=+4688.264362863" lastFinishedPulling="2025-12-03 22:59:55.160734817 +0000 UTC m=+4691.903445412" observedRunningTime="2025-12-03 22:59:55.591962184 +0000 UTC m=+4692.334672779" watchObservedRunningTime="2025-12-03 22:59:55.596676533 +0000 UTC m=+4692.339387148" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.659174 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-sctl9_a91ec3b7-705a-4500-be8e-e8c789a2c973/manager/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.704944 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-d5phr_b8cbfb12-17c2-46bf-baea-ee8ac16db4ee/kube-rbac-proxy/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.843912 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-d5phr_b8cbfb12-17c2-46bf-baea-ee8ac16db4ee/manager/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.937457 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gc2vn_1cba7fbc-0a66-45b4-b20d-33c81d407b96/kube-rbac-proxy/0.log" Dec 03 22:59:55 crc kubenswrapper[4715]: I1203 22:59:55.998472 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gc2vn_1cba7fbc-0a66-45b4-b20d-33c81d407b96/manager/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.166901 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-tgkbx_8219810d-511e-4cc8-a018-49f425b339c0/manager/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.207087 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-tgkbx_8219810d-511e-4cc8-a018-49f425b339c0/kube-rbac-proxy/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.372178 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-kwnlp_6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef/kube-rbac-proxy/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.495597 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-9qbjh_95b90127-8325-43c8-977e-8b972b62db2a/kube-rbac-proxy/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.537489 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-kwnlp_6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef/manager/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.617631 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-9qbjh_95b90127-8325-43c8-977e-8b972b62db2a/manager/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.747787 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qbp4s_305a708b-69fc-4263-a28f-015fe16402f7/kube-rbac-proxy/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.864791 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qbp4s_305a708b-69fc-4263-a28f-015fe16402f7/manager/0.log" Dec 03 22:59:56 crc kubenswrapper[4715]: I1203 22:59:56.968322 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-9gf6m_9a900ae2-08bc-403b-a613-52e7870eeebf/manager/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.004580 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-9gf6m_9a900ae2-08bc-403b-a613-52e7870eeebf/kube-rbac-proxy/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.069635 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hqgrk_c6d7646e-fb4a-4199-aadb-670000485477/kube-rbac-proxy/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.192586 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hqgrk_c6d7646e-fb4a-4199-aadb-670000485477/manager/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.305782 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-gv4jj_3091b003-a466-4d34-8a26-d3e68a3639a9/kube-rbac-proxy/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.353470 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-gv4jj_3091b003-a466-4d34-8a26-d3e68a3639a9/manager/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.454025 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-hm2q2_b8304681-abf0-4e5c-aa37-41453f6e7731/kube-rbac-proxy/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.613110 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-hm2q2_b8304681-abf0-4e5c-aa37-41453f6e7731/manager/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.685809 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-9ch2g_594f9dc7-fb50-4241-bb93-90d8039339dc/kube-rbac-proxy/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.719775 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-9ch2g_594f9dc7-fb50-4241-bb93-90d8039339dc/manager/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.875639 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp_dbeb21cb-a705-464d-ace4-90e1ce5084a5/kube-rbac-proxy/0.log" Dec 03 22:59:57 crc kubenswrapper[4715]: I1203 22:59:57.956699 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp_dbeb21cb-a705-464d-ace4-90e1ce5084a5/manager/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.220900 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tpfhd_dc2ba455-421c-4b30-9a6c-93945d7d1c9f/registry-server/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.305753 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-85f8f75cbc-4ld8x_d4195e77-3226-4e2f-b29e-b6ddaaa3bef2/operator/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.444925 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-68hlj_c7e152eb-ab3c-427a-a12b-70f46b881eee/kube-rbac-proxy/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.510932 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-hj6qn_6925aa7e-915e-4bbf-a3e1-991ce581b49c/kube-rbac-proxy/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.531577 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-68hlj_c7e152eb-ab3c-427a-a12b-70f46b881eee/manager/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.687243 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-hj6qn_6925aa7e-915e-4bbf-a3e1-991ce581b49c/manager/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.851980 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-hnzzx_1d2cfa37-465d-4b27-8a8b-90038e1f4bf1/operator/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.990917 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-252wh_f621de25-14cd-43f5-a0a8-cf0866337715/kube-rbac-proxy/0.log" Dec 03 22:59:58 crc kubenswrapper[4715]: I1203 22:59:58.993856 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-252wh_f621de25-14cd-43f5-a0a8-cf0866337715/manager/0.log" Dec 03 22:59:59 crc kubenswrapper[4715]: I1203 22:59:59.037112 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-bd9c944fb-mkhhs_a41a8afb-d557-4caa-bf3e-57a9a5115f57/manager/0.log" Dec 03 22:59:59 crc kubenswrapper[4715]: I1203 22:59:59.292826 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-8g9vt_85bafa11-dcf4-4ea4-b68e-e2d4799e813d/kube-rbac-proxy/0.log" Dec 03 22:59:59 crc kubenswrapper[4715]: I1203 22:59:59.393036 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-8g9vt_85bafa11-dcf4-4ea4-b68e-e2d4799e813d/manager/0.log" Dec 03 22:59:59 crc kubenswrapper[4715]: I1203 22:59:59.437264 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-ttfdp_0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e/kube-rbac-proxy/0.log" Dec 03 22:59:59 crc kubenswrapper[4715]: I1203 22:59:59.481423 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-ttfdp_0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e/manager/0.log" Dec 03 22:59:59 crc kubenswrapper[4715]: I1203 22:59:59.589210 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-gtpjt_959f935f-f997-47bc-8798-d54f7b62f2c9/kube-rbac-proxy/0.log" Dec 03 22:59:59 crc kubenswrapper[4715]: I1203 22:59:59.639640 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-gtpjt_959f935f-f997-47bc-8798-d54f7b62f2c9/manager/0.log" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.185894 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj"] Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.187116 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.192525 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.192534 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.198419 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj"] Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.223942 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zd62w" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.225405 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zd62w" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.271925 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zd62w" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.370224 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dc151e23-e695-4814-b191-7981042e077f-config-volume\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.370308 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dc151e23-e695-4814-b191-7981042e077f-secret-volume\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.370341 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9v9g\" (UniqueName: \"kubernetes.io/projected/dc151e23-e695-4814-b191-7981042e077f-kube-api-access-r9v9g\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.472078 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dc151e23-e695-4814-b191-7981042e077f-secret-volume\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.472151 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9v9g\" (UniqueName: \"kubernetes.io/projected/dc151e23-e695-4814-b191-7981042e077f-kube-api-access-r9v9g\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.472346 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dc151e23-e695-4814-b191-7981042e077f-config-volume\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.473459 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dc151e23-e695-4814-b191-7981042e077f-config-volume\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.478974 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dc151e23-e695-4814-b191-7981042e077f-secret-volume\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.492978 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9v9g\" (UniqueName: \"kubernetes.io/projected/dc151e23-e695-4814-b191-7981042e077f-kube-api-access-r9v9g\") pod \"collect-profiles-29413380-ntpcj\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.516808 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.683363 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zd62w" Dec 03 23:00:00 crc kubenswrapper[4715]: I1203 23:00:00.777734 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zd62w"] Dec 03 23:00:01 crc kubenswrapper[4715]: I1203 23:00:01.019759 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj"] Dec 03 23:00:01 crc kubenswrapper[4715]: W1203 23:00:01.024868 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc151e23_e695_4814_b191_7981042e077f.slice/crio-8dabf325999e94466671b5f1a15989acef8f82a881f791d49ef8fe9e58f6d0d8 WatchSource:0}: Error finding container 8dabf325999e94466671b5f1a15989acef8f82a881f791d49ef8fe9e58f6d0d8: Status 404 returned error can't find the container with id 8dabf325999e94466671b5f1a15989acef8f82a881f791d49ef8fe9e58f6d0d8 Dec 03 23:00:01 crc kubenswrapper[4715]: I1203 23:00:01.632417 4715 generic.go:334] "Generic (PLEG): container finished" podID="dc151e23-e695-4814-b191-7981042e077f" containerID="72135f8bb901bf38bfe5addbae13922432b6f8b3d411fa7b288838062e6f59e4" exitCode=0 Dec 03 23:00:01 crc kubenswrapper[4715]: I1203 23:00:01.632630 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" event={"ID":"dc151e23-e695-4814-b191-7981042e077f","Type":"ContainerDied","Data":"72135f8bb901bf38bfe5addbae13922432b6f8b3d411fa7b288838062e6f59e4"} Dec 03 23:00:01 crc kubenswrapper[4715]: I1203 23:00:01.645133 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" event={"ID":"dc151e23-e695-4814-b191-7981042e077f","Type":"ContainerStarted","Data":"8dabf325999e94466671b5f1a15989acef8f82a881f791d49ef8fe9e58f6d0d8"} Dec 03 23:00:02 crc kubenswrapper[4715]: I1203 23:00:02.641209 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zd62w" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerName="registry-server" containerID="cri-o://c1642f36dd288afa37f27a2367b9230b08ce5b1e332c93ad4648b57e15bb7971" gracePeriod=2 Dec 03 23:00:02 crc kubenswrapper[4715]: I1203 23:00:02.984833 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.122117 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dc151e23-e695-4814-b191-7981042e077f-config-volume\") pod \"dc151e23-e695-4814-b191-7981042e077f\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.122311 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9v9g\" (UniqueName: \"kubernetes.io/projected/dc151e23-e695-4814-b191-7981042e077f-kube-api-access-r9v9g\") pod \"dc151e23-e695-4814-b191-7981042e077f\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.122408 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dc151e23-e695-4814-b191-7981042e077f-secret-volume\") pod \"dc151e23-e695-4814-b191-7981042e077f\" (UID: \"dc151e23-e695-4814-b191-7981042e077f\") " Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.123028 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc151e23-e695-4814-b191-7981042e077f-config-volume" (OuterVolumeSpecName: "config-volume") pod "dc151e23-e695-4814-b191-7981042e077f" (UID: "dc151e23-e695-4814-b191-7981042e077f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.129060 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc151e23-e695-4814-b191-7981042e077f-kube-api-access-r9v9g" (OuterVolumeSpecName: "kube-api-access-r9v9g") pod "dc151e23-e695-4814-b191-7981042e077f" (UID: "dc151e23-e695-4814-b191-7981042e077f"). InnerVolumeSpecName "kube-api-access-r9v9g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.129321 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc151e23-e695-4814-b191-7981042e077f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dc151e23-e695-4814-b191-7981042e077f" (UID: "dc151e23-e695-4814-b191-7981042e077f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.225177 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dc151e23-e695-4814-b191-7981042e077f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.225219 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9v9g\" (UniqueName: \"kubernetes.io/projected/dc151e23-e695-4814-b191-7981042e077f-kube-api-access-r9v9g\") on node \"crc\" DevicePath \"\"" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.225233 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dc151e23-e695-4814-b191-7981042e077f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.652872 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.652881 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" event={"ID":"dc151e23-e695-4814-b191-7981042e077f","Type":"ContainerDied","Data":"8dabf325999e94466671b5f1a15989acef8f82a881f791d49ef8fe9e58f6d0d8"} Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.652921 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dabf325999e94466671b5f1a15989acef8f82a881f791d49ef8fe9e58f6d0d8" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.660730 4715 generic.go:334] "Generic (PLEG): container finished" podID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerID="c1642f36dd288afa37f27a2367b9230b08ce5b1e332c93ad4648b57e15bb7971" exitCode=0 Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.660778 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd62w" event={"ID":"561f5f6c-67b2-4d46-b595-339ad5137f48","Type":"ContainerDied","Data":"c1642f36dd288afa37f27a2367b9230b08ce5b1e332c93ad4648b57e15bb7971"} Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.660807 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zd62w" event={"ID":"561f5f6c-67b2-4d46-b595-339ad5137f48","Type":"ContainerDied","Data":"7e348136f6c4b8f279cbff0b378a696a96440574fa4b84cf910d9a6e851dd47e"} Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.660820 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e348136f6c4b8f279cbff0b378a696a96440574fa4b84cf910d9a6e851dd47e" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.676048 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zd62w" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.834434 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-catalog-content\") pod \"561f5f6c-67b2-4d46-b595-339ad5137f48\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.834542 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-utilities\") pod \"561f5f6c-67b2-4d46-b595-339ad5137f48\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.834694 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p494r\" (UniqueName: \"kubernetes.io/projected/561f5f6c-67b2-4d46-b595-339ad5137f48-kube-api-access-p494r\") pod \"561f5f6c-67b2-4d46-b595-339ad5137f48\" (UID: \"561f5f6c-67b2-4d46-b595-339ad5137f48\") " Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.839593 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-utilities" (OuterVolumeSpecName: "utilities") pod "561f5f6c-67b2-4d46-b595-339ad5137f48" (UID: "561f5f6c-67b2-4d46-b595-339ad5137f48"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.845394 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561f5f6c-67b2-4d46-b595-339ad5137f48-kube-api-access-p494r" (OuterVolumeSpecName: "kube-api-access-p494r") pod "561f5f6c-67b2-4d46-b595-339ad5137f48" (UID: "561f5f6c-67b2-4d46-b595-339ad5137f48"). InnerVolumeSpecName "kube-api-access-p494r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.897338 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "561f5f6c-67b2-4d46-b595-339ad5137f48" (UID: "561f5f6c-67b2-4d46-b595-339ad5137f48"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.940145 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.940180 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561f5f6c-67b2-4d46-b595-339ad5137f48-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 23:00:03 crc kubenswrapper[4715]: I1203 23:00:03.940189 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p494r\" (UniqueName: \"kubernetes.io/projected/561f5f6c-67b2-4d46-b595-339ad5137f48-kube-api-access-p494r\") on node \"crc\" DevicePath \"\"" Dec 03 23:00:04 crc kubenswrapper[4715]: I1203 23:00:04.072048 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2"] Dec 03 23:00:04 crc kubenswrapper[4715]: I1203 23:00:04.084435 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413335-5pbl2"] Dec 03 23:00:04 crc kubenswrapper[4715]: I1203 23:00:04.669621 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zd62w" Dec 03 23:00:04 crc kubenswrapper[4715]: I1203 23:00:04.722309 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zd62w"] Dec 03 23:00:04 crc kubenswrapper[4715]: I1203 23:00:04.736443 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zd62w"] Dec 03 23:00:05 crc kubenswrapper[4715]: I1203 23:00:05.646085 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" path="/var/lib/kubelet/pods/561f5f6c-67b2-4d46-b595-339ad5137f48/volumes" Dec 03 23:00:05 crc kubenswrapper[4715]: I1203 23:00:05.647337 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6579b54-5da2-431c-beb6-e688bc93560c" path="/var/lib/kubelet/pods/e6579b54-5da2-431c-beb6-e688bc93560c/volumes" Dec 03 23:00:18 crc kubenswrapper[4715]: I1203 23:00:18.568107 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dwjvx_290986b1-4ac0-40df-9294-0cff8d471733/control-plane-machine-set-operator/0.log" Dec 03 23:00:18 crc kubenswrapper[4715]: I1203 23:00:18.745346 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-77n4d_cd7d035c-0bd3-4568-8384-03b463fb0c5c/kube-rbac-proxy/0.log" Dec 03 23:00:18 crc kubenswrapper[4715]: I1203 23:00:18.836239 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-77n4d_cd7d035c-0bd3-4568-8384-03b463fb0c5c/machine-api-operator/0.log" Dec 03 23:00:31 crc kubenswrapper[4715]: I1203 23:00:31.148037 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-4bvpr_ce5706ce-67d9-4e61-8051-03113e8c3ac4/cert-manager-controller/0.log" Dec 03 23:00:31 crc kubenswrapper[4715]: I1203 23:00:31.296827 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rnccs_bec62745-cf3a-4bc6-9fc6-34c8a56301ce/cert-manager-cainjector/0.log" Dec 03 23:00:31 crc kubenswrapper[4715]: I1203 23:00:31.383582 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dgskj_30ac245b-49dd-48b9-878e-e06008f921b7/cert-manager-webhook/0.log" Dec 03 23:00:33 crc kubenswrapper[4715]: I1203 23:00:33.671714 4715 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","poddc151e23-e695-4814-b191-7981042e077f"] err="unable to destroy cgroup paths for cgroup [kubepods burstable poddc151e23-e695-4814-b191-7981042e077f] : Timed out while waiting for systemd to remove kubepods-burstable-poddc151e23_e695_4814_b191_7981042e077f.slice" Dec 03 23:00:33 crc kubenswrapper[4715]: E1203 23:00:33.672248 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods burstable poddc151e23-e695-4814-b191-7981042e077f] : unable to destroy cgroup paths for cgroup [kubepods burstable poddc151e23-e695-4814-b191-7981042e077f] : Timed out while waiting for systemd to remove kubepods-burstable-poddc151e23_e695_4814_b191_7981042e077f.slice" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" podUID="dc151e23-e695-4814-b191-7981042e077f" Dec 03 23:00:33 crc kubenswrapper[4715]: I1203 23:00:33.909788 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413380-ntpcj" Dec 03 23:00:42 crc kubenswrapper[4715]: I1203 23:00:42.801333 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-lkqpr_47af66c3-ee08-4f79-9e8e-f9e1b3457971/nmstate-console-plugin/0.log" Dec 03 23:00:42 crc kubenswrapper[4715]: I1203 23:00:42.964322 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vkmgw_be31b811-7008-4d9f-ac8b-4c454e4a8597/nmstate-handler/0.log" Dec 03 23:00:43 crc kubenswrapper[4715]: I1203 23:00:43.031607 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9n62x_6d716f67-d1d1-44c0-b2d4-eeb60e73160f/nmstate-metrics/0.log" Dec 03 23:00:43 crc kubenswrapper[4715]: I1203 23:00:43.057307 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9n62x_6d716f67-d1d1-44c0-b2d4-eeb60e73160f/kube-rbac-proxy/0.log" Dec 03 23:00:43 crc kubenswrapper[4715]: I1203 23:00:43.264089 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-xht5s_7bff12a8-00c5-4a84-85f7-3a0055308b87/nmstate-operator/0.log" Dec 03 23:00:43 crc kubenswrapper[4715]: I1203 23:00:43.285627 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-4d8n9_bef574d3-8639-4a82-aa6e-323a58b9db40/nmstate-webhook/0.log" Dec 03 23:00:56 crc kubenswrapper[4715]: I1203 23:00:56.621583 4715 scope.go:117] "RemoveContainer" containerID="c07ff48571ffc753f77a1e86664859a3317ec1ca39f20791fa1aa3d1cbf323b6" Dec 03 23:00:57 crc kubenswrapper[4715]: I1203 23:00:57.155273 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-22qzs_a7a0c0e5-fb39-479e-8731-f81bc225c7fb/kube-rbac-proxy/0.log" Dec 03 23:00:57 crc kubenswrapper[4715]: I1203 23:00:57.318527 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-22qzs_a7a0c0e5-fb39-479e-8731-f81bc225c7fb/controller/0.log" Dec 03 23:00:57 crc kubenswrapper[4715]: I1203 23:00:57.417232 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-frr-files/0.log" Dec 03 23:00:57 crc kubenswrapper[4715]: I1203 23:00:57.566052 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-reloader/0.log" Dec 03 23:00:57 crc kubenswrapper[4715]: I1203 23:00:57.592431 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-frr-files/0.log" Dec 03 23:00:57 crc kubenswrapper[4715]: I1203 23:00:57.611330 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-reloader/0.log" Dec 03 23:00:57 crc kubenswrapper[4715]: I1203 23:00:57.611561 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-metrics/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.002407 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-metrics/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.006417 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-frr-files/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.035946 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-metrics/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.066321 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-reloader/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.203146 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-frr-files/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.214197 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-reloader/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.243798 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/controller/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.287221 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-metrics/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.368390 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/frr-metrics/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.488900 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/kube-rbac-proxy-frr/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.549084 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/kube-rbac-proxy/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.630618 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/reloader/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.783837 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-94tb9_ff247a17-7b16-4a5e-adcf-48f9122130eb/frr-k8s-webhook-server/0.log" Dec 03 23:00:58 crc kubenswrapper[4715]: I1203 23:00:58.935003 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-b57df49df-j4jvw_3483831a-11e2-4881-8447-fc127b2ec983/manager/0.log" Dec 03 23:00:59 crc kubenswrapper[4715]: I1203 23:00:59.153604 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7f9b456564-mq9gn_0b23c7d6-56da-4a46-ab0e-e98665166baa/webhook-server/0.log" Dec 03 23:00:59 crc kubenswrapper[4715]: I1203 23:00:59.290153 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rjg8s_b396984d-e5db-4417-ab0b-c2846cc8c95a/kube-rbac-proxy/0.log" Dec 03 23:00:59 crc kubenswrapper[4715]: I1203 23:00:59.899572 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rjg8s_b396984d-e5db-4417-ab0b-c2846cc8c95a/speaker/0.log" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.079558 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/frr/0.log" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.156402 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29413381-ncq8v"] Dec 03 23:01:00 crc kubenswrapper[4715]: E1203 23:01:00.156898 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerName="extract-content" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.156913 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerName="extract-content" Dec 03 23:01:00 crc kubenswrapper[4715]: E1203 23:01:00.156923 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerName="registry-server" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.156929 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerName="registry-server" Dec 03 23:01:00 crc kubenswrapper[4715]: E1203 23:01:00.156945 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerName="extract-utilities" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.156951 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerName="extract-utilities" Dec 03 23:01:00 crc kubenswrapper[4715]: E1203 23:01:00.156963 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc151e23-e695-4814-b191-7981042e077f" containerName="collect-profiles" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.156969 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc151e23-e695-4814-b191-7981042e077f" containerName="collect-profiles" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.157146 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="561f5f6c-67b2-4d46-b595-339ad5137f48" containerName="registry-server" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.157160 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc151e23-e695-4814-b191-7981042e077f" containerName="collect-profiles" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.157872 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.165144 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413381-ncq8v"] Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.258395 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-combined-ca-bundle\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.258462 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-config-data\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.258484 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpfnw\" (UniqueName: \"kubernetes.io/projected/a2c5df74-94f6-4aa7-a66a-386414a454ce-kube-api-access-fpfnw\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.258528 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-fernet-keys\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.360984 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-combined-ca-bundle\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.361055 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-config-data\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.361086 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpfnw\" (UniqueName: \"kubernetes.io/projected/a2c5df74-94f6-4aa7-a66a-386414a454ce-kube-api-access-fpfnw\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.361118 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-fernet-keys\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.367795 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-combined-ca-bundle\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.369562 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-fernet-keys\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.369618 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-config-data\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.396064 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpfnw\" (UniqueName: \"kubernetes.io/projected/a2c5df74-94f6-4aa7-a66a-386414a454ce-kube-api-access-fpfnw\") pod \"keystone-cron-29413381-ncq8v\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.499741 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:00 crc kubenswrapper[4715]: I1203 23:01:00.993683 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413381-ncq8v"] Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.151227 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413381-ncq8v" event={"ID":"a2c5df74-94f6-4aa7-a66a-386414a454ce","Type":"ContainerStarted","Data":"23bddbba8378151d5e541c5459bee0819cdee414e29cc4af963a7e8c42151d68"} Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.311425 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pqgwq"] Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.314177 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.343068 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqgwq"] Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.384656 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx2n8\" (UniqueName: \"kubernetes.io/projected/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-kube-api-access-gx2n8\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.384733 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-catalog-content\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.384987 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-utilities\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.486692 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx2n8\" (UniqueName: \"kubernetes.io/projected/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-kube-api-access-gx2n8\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.486740 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-catalog-content\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.486886 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-utilities\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.487486 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-utilities\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.487589 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-catalog-content\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.517325 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx2n8\" (UniqueName: \"kubernetes.io/projected/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-kube-api-access-gx2n8\") pod \"redhat-marketplace-pqgwq\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:01 crc kubenswrapper[4715]: I1203 23:01:01.641780 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:02 crc kubenswrapper[4715]: I1203 23:01:02.111863 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqgwq"] Dec 03 23:01:02 crc kubenswrapper[4715]: W1203 23:01:02.120346 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode50bec72_ed8a_4f57_a8a0_3b5bf9164c09.slice/crio-781c248eda1fa31c3a5d1e1e82849817aaf836c1cc49dcf7e4b72e5478f59aa9 WatchSource:0}: Error finding container 781c248eda1fa31c3a5d1e1e82849817aaf836c1cc49dcf7e4b72e5478f59aa9: Status 404 returned error can't find the container with id 781c248eda1fa31c3a5d1e1e82849817aaf836c1cc49dcf7e4b72e5478f59aa9 Dec 03 23:01:02 crc kubenswrapper[4715]: I1203 23:01:02.170098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqgwq" event={"ID":"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09","Type":"ContainerStarted","Data":"781c248eda1fa31c3a5d1e1e82849817aaf836c1cc49dcf7e4b72e5478f59aa9"} Dec 03 23:01:02 crc kubenswrapper[4715]: I1203 23:01:02.171877 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413381-ncq8v" event={"ID":"a2c5df74-94f6-4aa7-a66a-386414a454ce","Type":"ContainerStarted","Data":"906ded87027af954adcc4ebe250869c324780ee63f3d3dada19dad0f6606d51e"} Dec 03 23:01:02 crc kubenswrapper[4715]: I1203 23:01:02.192888 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29413381-ncq8v" podStartSLOduration=2.192869727 podStartE2EDuration="2.192869727s" podCreationTimestamp="2025-12-03 23:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 23:01:02.19177757 +0000 UTC m=+4758.934488165" watchObservedRunningTime="2025-12-03 23:01:02.192869727 +0000 UTC m=+4758.935580332" Dec 03 23:01:03 crc kubenswrapper[4715]: I1203 23:01:03.183124 4715 generic.go:334] "Generic (PLEG): container finished" podID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerID="650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d" exitCode=0 Dec 03 23:01:03 crc kubenswrapper[4715]: I1203 23:01:03.183180 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqgwq" event={"ID":"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09","Type":"ContainerDied","Data":"650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d"} Dec 03 23:01:04 crc kubenswrapper[4715]: I1203 23:01:04.193252 4715 generic.go:334] "Generic (PLEG): container finished" podID="a2c5df74-94f6-4aa7-a66a-386414a454ce" containerID="906ded87027af954adcc4ebe250869c324780ee63f3d3dada19dad0f6606d51e" exitCode=0 Dec 03 23:01:04 crc kubenswrapper[4715]: I1203 23:01:04.193527 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413381-ncq8v" event={"ID":"a2c5df74-94f6-4aa7-a66a-386414a454ce","Type":"ContainerDied","Data":"906ded87027af954adcc4ebe250869c324780ee63f3d3dada19dad0f6606d51e"} Dec 03 23:01:04 crc kubenswrapper[4715]: I1203 23:01:04.195437 4715 generic.go:334] "Generic (PLEG): container finished" podID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerID="b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3" exitCode=0 Dec 03 23:01:04 crc kubenswrapper[4715]: I1203 23:01:04.195461 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqgwq" event={"ID":"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09","Type":"ContainerDied","Data":"b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3"} Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.160051 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.160428 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.208291 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqgwq" event={"ID":"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09","Type":"ContainerStarted","Data":"5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457"} Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.239776 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pqgwq" podStartSLOduration=2.8190776829999997 podStartE2EDuration="4.239754433s" podCreationTimestamp="2025-12-03 23:01:01 +0000 UTC" firstStartedPulling="2025-12-03 23:01:03.185370746 +0000 UTC m=+4759.928081331" lastFinishedPulling="2025-12-03 23:01:04.606047486 +0000 UTC m=+4761.348758081" observedRunningTime="2025-12-03 23:01:05.229087394 +0000 UTC m=+4761.971797989" watchObservedRunningTime="2025-12-03 23:01:05.239754433 +0000 UTC m=+4761.982465018" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.565110 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.662107 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-fernet-keys\") pod \"a2c5df74-94f6-4aa7-a66a-386414a454ce\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.662153 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-combined-ca-bundle\") pod \"a2c5df74-94f6-4aa7-a66a-386414a454ce\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.662277 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpfnw\" (UniqueName: \"kubernetes.io/projected/a2c5df74-94f6-4aa7-a66a-386414a454ce-kube-api-access-fpfnw\") pod \"a2c5df74-94f6-4aa7-a66a-386414a454ce\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.662461 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-config-data\") pod \"a2c5df74-94f6-4aa7-a66a-386414a454ce\" (UID: \"a2c5df74-94f6-4aa7-a66a-386414a454ce\") " Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.669450 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a2c5df74-94f6-4aa7-a66a-386414a454ce" (UID: "a2c5df74-94f6-4aa7-a66a-386414a454ce"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.669450 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2c5df74-94f6-4aa7-a66a-386414a454ce-kube-api-access-fpfnw" (OuterVolumeSpecName: "kube-api-access-fpfnw") pod "a2c5df74-94f6-4aa7-a66a-386414a454ce" (UID: "a2c5df74-94f6-4aa7-a66a-386414a454ce"). InnerVolumeSpecName "kube-api-access-fpfnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.691147 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2c5df74-94f6-4aa7-a66a-386414a454ce" (UID: "a2c5df74-94f6-4aa7-a66a-386414a454ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.718317 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-config-data" (OuterVolumeSpecName: "config-data") pod "a2c5df74-94f6-4aa7-a66a-386414a454ce" (UID: "a2c5df74-94f6-4aa7-a66a-386414a454ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.765020 4715 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.765057 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpfnw\" (UniqueName: \"kubernetes.io/projected/a2c5df74-94f6-4aa7-a66a-386414a454ce-kube-api-access-fpfnw\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.765069 4715 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:05 crc kubenswrapper[4715]: I1203 23:01:05.765078 4715 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a2c5df74-94f6-4aa7-a66a-386414a454ce-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:06 crc kubenswrapper[4715]: I1203 23:01:06.219240 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413381-ncq8v" event={"ID":"a2c5df74-94f6-4aa7-a66a-386414a454ce","Type":"ContainerDied","Data":"23bddbba8378151d5e541c5459bee0819cdee414e29cc4af963a7e8c42151d68"} Dec 03 23:01:06 crc kubenswrapper[4715]: I1203 23:01:06.219293 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23bddbba8378151d5e541c5459bee0819cdee414e29cc4af963a7e8c42151d68" Dec 03 23:01:06 crc kubenswrapper[4715]: I1203 23:01:06.219393 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413381-ncq8v" Dec 03 23:01:11 crc kubenswrapper[4715]: I1203 23:01:11.647774 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:11 crc kubenswrapper[4715]: I1203 23:01:11.648376 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:11 crc kubenswrapper[4715]: I1203 23:01:11.694687 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.253522 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/util/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.322472 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.376525 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqgwq"] Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.433567 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/util/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.472716 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/pull/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.475772 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/pull/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.608832 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/util/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.639896 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/pull/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.663009 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/extract/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.805705 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/util/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.960111 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/util/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.979645 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/pull/0.log" Dec 03 23:01:12 crc kubenswrapper[4715]: I1203 23:01:12.986275 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/pull/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.175038 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/extract/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.181743 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/util/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.197266 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/pull/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.324328 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-utilities/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.484282 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-utilities/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.496404 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-content/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.496412 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-content/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.654094 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-content/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.683246 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-utilities/0.log" Dec 03 23:01:13 crc kubenswrapper[4715]: I1203 23:01:13.841237 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-utilities/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.076423 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-content/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.111700 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-content/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.131100 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-utilities/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.291823 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pqgwq" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerName="registry-server" containerID="cri-o://5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457" gracePeriod=2 Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.303205 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/registry-server/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.346768 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-utilities/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.359156 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-content/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.637227 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ss8p4_62dffbaa-068c-484d-822d-51bb3bfd3105/marketplace-operator/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.779601 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.839058 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-catalog-content\") pod \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.839435 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-utilities\") pod \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.839465 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx2n8\" (UniqueName: \"kubernetes.io/projected/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-kube-api-access-gx2n8\") pod \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\" (UID: \"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09\") " Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.845118 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-utilities" (OuterVolumeSpecName: "utilities") pod "e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" (UID: "e50bec72-ed8a-4f57-a8a0-3b5bf9164c09"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.846223 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pqgwq_e50bec72-ed8a-4f57-a8a0-3b5bf9164c09/extract-utilities/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.850296 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-kube-api-access-gx2n8" (OuterVolumeSpecName: "kube-api-access-gx2n8") pod "e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" (UID: "e50bec72-ed8a-4f57-a8a0-3b5bf9164c09"). InnerVolumeSpecName "kube-api-access-gx2n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.873325 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" (UID: "e50bec72-ed8a-4f57-a8a0-3b5bf9164c09"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.911175 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/registry-server/0.log" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.941695 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.941730 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx2n8\" (UniqueName: \"kubernetes.io/projected/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-kube-api-access-gx2n8\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:14 crc kubenswrapper[4715]: I1203 23:01:14.941741 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.021689 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pqgwq_e50bec72-ed8a-4f57-a8a0-3b5bf9164c09/extract-utilities/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.035811 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pqgwq_e50bec72-ed8a-4f57-a8a0-3b5bf9164c09/extract-content/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.063892 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pqgwq_e50bec72-ed8a-4f57-a8a0-3b5bf9164c09/extract-content/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.255046 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pqgwq_e50bec72-ed8a-4f57-a8a0-3b5bf9164c09/extract-content/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.266608 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pqgwq_e50bec72-ed8a-4f57-a8a0-3b5bf9164c09/extract-utilities/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.279464 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-pqgwq_e50bec72-ed8a-4f57-a8a0-3b5bf9164c09/registry-server/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.301771 4715 generic.go:334] "Generic (PLEG): container finished" podID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerID="5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457" exitCode=0 Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.301819 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqgwq" event={"ID":"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09","Type":"ContainerDied","Data":"5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457"} Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.301854 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pqgwq" event={"ID":"e50bec72-ed8a-4f57-a8a0-3b5bf9164c09","Type":"ContainerDied","Data":"781c248eda1fa31c3a5d1e1e82849817aaf836c1cc49dcf7e4b72e5478f59aa9"} Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.301863 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pqgwq" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.301875 4715 scope.go:117] "RemoveContainer" containerID="5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.324951 4715 scope.go:117] "RemoveContainer" containerID="b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.341683 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqgwq"] Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.349980 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pqgwq"] Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.363385 4715 scope.go:117] "RemoveContainer" containerID="650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.398734 4715 scope.go:117] "RemoveContainer" containerID="5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457" Dec 03 23:01:15 crc kubenswrapper[4715]: E1203 23:01:15.399222 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457\": container with ID starting with 5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457 not found: ID does not exist" containerID="5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.399258 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457"} err="failed to get container status \"5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457\": rpc error: code = NotFound desc = could not find container \"5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457\": container with ID starting with 5bb64b1567ba92b433178076c0a903d1bd988f03d72f693b796cdee0fdab3457 not found: ID does not exist" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.399288 4715 scope.go:117] "RemoveContainer" containerID="b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3" Dec 03 23:01:15 crc kubenswrapper[4715]: E1203 23:01:15.399747 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3\": container with ID starting with b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3 not found: ID does not exist" containerID="b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.399772 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3"} err="failed to get container status \"b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3\": rpc error: code = NotFound desc = could not find container \"b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3\": container with ID starting with b6bb64a66728f933da9409d35a69274cf4ccefb4201d945e44fa5b1c8baa39f3 not found: ID does not exist" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.399803 4715 scope.go:117] "RemoveContainer" containerID="650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d" Dec 03 23:01:15 crc kubenswrapper[4715]: E1203 23:01:15.400216 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d\": container with ID starting with 650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d not found: ID does not exist" containerID="650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.400237 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d"} err="failed to get container status \"650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d\": rpc error: code = NotFound desc = could not find container \"650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d\": container with ID starting with 650637d76c0426c3ed4540a06d9edd8dc5ddf5488018fd40bb1192d57dc8fc8d not found: ID does not exist" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.435303 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-utilities/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.646997 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" path="/var/lib/kubelet/pods/e50bec72-ed8a-4f57-a8a0-3b5bf9164c09/volumes" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.658612 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-content/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.690728 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-utilities/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.719891 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-content/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.876182 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-content/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.918059 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-utilities/0.log" Dec 03 23:01:15 crc kubenswrapper[4715]: I1203 23:01:15.931724 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-utilities/0.log" Dec 03 23:01:16 crc kubenswrapper[4715]: I1203 23:01:16.007416 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/registry-server/0.log" Dec 03 23:01:16 crc kubenswrapper[4715]: I1203 23:01:16.138824 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-content/0.log" Dec 03 23:01:16 crc kubenswrapper[4715]: I1203 23:01:16.144153 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-utilities/0.log" Dec 03 23:01:16 crc kubenswrapper[4715]: I1203 23:01:16.145776 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-content/0.log" Dec 03 23:01:16 crc kubenswrapper[4715]: I1203 23:01:16.313599 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-content/0.log" Dec 03 23:01:16 crc kubenswrapper[4715]: I1203 23:01:16.336446 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-utilities/0.log" Dec 03 23:01:16 crc kubenswrapper[4715]: I1203 23:01:16.908700 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/registry-server/0.log" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.205745 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6nms2"] Dec 03 23:01:25 crc kubenswrapper[4715]: E1203 23:01:25.206806 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerName="extract-content" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.206825 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerName="extract-content" Dec 03 23:01:25 crc kubenswrapper[4715]: E1203 23:01:25.206849 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerName="extract-utilities" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.206858 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerName="extract-utilities" Dec 03 23:01:25 crc kubenswrapper[4715]: E1203 23:01:25.206881 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2c5df74-94f6-4aa7-a66a-386414a454ce" containerName="keystone-cron" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.206889 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2c5df74-94f6-4aa7-a66a-386414a454ce" containerName="keystone-cron" Dec 03 23:01:25 crc kubenswrapper[4715]: E1203 23:01:25.206923 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerName="registry-server" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.206931 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerName="registry-server" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.207177 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2c5df74-94f6-4aa7-a66a-386414a454ce" containerName="keystone-cron" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.207202 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50bec72-ed8a-4f57-a8a0-3b5bf9164c09" containerName="registry-server" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.208898 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.218287 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nms2"] Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.231364 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-utilities\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.231418 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4qck\" (UniqueName: \"kubernetes.io/projected/011252d1-7f62-447b-89d3-134e1ade1471-kube-api-access-k4qck\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.231528 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-catalog-content\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.334471 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-utilities\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.334564 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4qck\" (UniqueName: \"kubernetes.io/projected/011252d1-7f62-447b-89d3-134e1ade1471-kube-api-access-k4qck\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.334656 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-catalog-content\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.335208 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-catalog-content\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.335207 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-utilities\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.357192 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4qck\" (UniqueName: \"kubernetes.io/projected/011252d1-7f62-447b-89d3-134e1ade1471-kube-api-access-k4qck\") pod \"community-operators-6nms2\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:25 crc kubenswrapper[4715]: I1203 23:01:25.530314 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:26 crc kubenswrapper[4715]: I1203 23:01:26.071862 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6nms2"] Dec 03 23:01:26 crc kubenswrapper[4715]: I1203 23:01:26.427567 4715 generic.go:334] "Generic (PLEG): container finished" podID="011252d1-7f62-447b-89d3-134e1ade1471" containerID="07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877" exitCode=0 Dec 03 23:01:26 crc kubenswrapper[4715]: I1203 23:01:26.427838 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nms2" event={"ID":"011252d1-7f62-447b-89d3-134e1ade1471","Type":"ContainerDied","Data":"07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877"} Dec 03 23:01:26 crc kubenswrapper[4715]: I1203 23:01:26.427871 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nms2" event={"ID":"011252d1-7f62-447b-89d3-134e1ade1471","Type":"ContainerStarted","Data":"925475722fd7619d172a68241422c7fe49cf6a43ae6aa8453c4a3e588fff3f04"} Dec 03 23:01:27 crc kubenswrapper[4715]: I1203 23:01:27.439060 4715 generic.go:334] "Generic (PLEG): container finished" podID="011252d1-7f62-447b-89d3-134e1ade1471" containerID="dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7" exitCode=0 Dec 03 23:01:27 crc kubenswrapper[4715]: I1203 23:01:27.439237 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nms2" event={"ID":"011252d1-7f62-447b-89d3-134e1ade1471","Type":"ContainerDied","Data":"dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7"} Dec 03 23:01:28 crc kubenswrapper[4715]: I1203 23:01:28.452146 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nms2" event={"ID":"011252d1-7f62-447b-89d3-134e1ade1471","Type":"ContainerStarted","Data":"3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18"} Dec 03 23:01:28 crc kubenswrapper[4715]: I1203 23:01:28.469146 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6nms2" podStartSLOduration=2.006306247 podStartE2EDuration="3.469127211s" podCreationTimestamp="2025-12-03 23:01:25 +0000 UTC" firstStartedPulling="2025-12-03 23:01:26.430080917 +0000 UTC m=+4783.172791512" lastFinishedPulling="2025-12-03 23:01:27.892901871 +0000 UTC m=+4784.635612476" observedRunningTime="2025-12-03 23:01:28.467468637 +0000 UTC m=+4785.210179233" watchObservedRunningTime="2025-12-03 23:01:28.469127211 +0000 UTC m=+4785.211837796" Dec 03 23:01:35 crc kubenswrapper[4715]: I1203 23:01:35.159872 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:01:35 crc kubenswrapper[4715]: I1203 23:01:35.160472 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:01:35 crc kubenswrapper[4715]: I1203 23:01:35.531296 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:35 crc kubenswrapper[4715]: I1203 23:01:35.531473 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:35 crc kubenswrapper[4715]: I1203 23:01:35.587365 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:36 crc kubenswrapper[4715]: I1203 23:01:36.584429 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:36 crc kubenswrapper[4715]: I1203 23:01:36.633444 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nms2"] Dec 03 23:01:38 crc kubenswrapper[4715]: I1203 23:01:38.553278 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6nms2" podUID="011252d1-7f62-447b-89d3-134e1ade1471" containerName="registry-server" containerID="cri-o://3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18" gracePeriod=2 Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.016913 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.119343 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-utilities\") pod \"011252d1-7f62-447b-89d3-134e1ade1471\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.119413 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-catalog-content\") pod \"011252d1-7f62-447b-89d3-134e1ade1471\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.119450 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4qck\" (UniqueName: \"kubernetes.io/projected/011252d1-7f62-447b-89d3-134e1ade1471-kube-api-access-k4qck\") pod \"011252d1-7f62-447b-89d3-134e1ade1471\" (UID: \"011252d1-7f62-447b-89d3-134e1ade1471\") " Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.119963 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-utilities" (OuterVolumeSpecName: "utilities") pod "011252d1-7f62-447b-89d3-134e1ade1471" (UID: "011252d1-7f62-447b-89d3-134e1ade1471"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.120277 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.151709 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/011252d1-7f62-447b-89d3-134e1ade1471-kube-api-access-k4qck" (OuterVolumeSpecName: "kube-api-access-k4qck") pod "011252d1-7f62-447b-89d3-134e1ade1471" (UID: "011252d1-7f62-447b-89d3-134e1ade1471"). InnerVolumeSpecName "kube-api-access-k4qck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.187639 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "011252d1-7f62-447b-89d3-134e1ade1471" (UID: "011252d1-7f62-447b-89d3-134e1ade1471"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.222128 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/011252d1-7f62-447b-89d3-134e1ade1471-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.222391 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4qck\" (UniqueName: \"kubernetes.io/projected/011252d1-7f62-447b-89d3-134e1ade1471-kube-api-access-k4qck\") on node \"crc\" DevicePath \"\"" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.570600 4715 generic.go:334] "Generic (PLEG): container finished" podID="011252d1-7f62-447b-89d3-134e1ade1471" containerID="3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18" exitCode=0 Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.570689 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6nms2" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.570714 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nms2" event={"ID":"011252d1-7f62-447b-89d3-134e1ade1471","Type":"ContainerDied","Data":"3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18"} Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.572029 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6nms2" event={"ID":"011252d1-7f62-447b-89d3-134e1ade1471","Type":"ContainerDied","Data":"925475722fd7619d172a68241422c7fe49cf6a43ae6aa8453c4a3e588fff3f04"} Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.572068 4715 scope.go:117] "RemoveContainer" containerID="3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.611220 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6nms2"] Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.617325 4715 scope.go:117] "RemoveContainer" containerID="dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.618086 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6nms2"] Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.641091 4715 scope.go:117] "RemoveContainer" containerID="07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.644023 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="011252d1-7f62-447b-89d3-134e1ade1471" path="/var/lib/kubelet/pods/011252d1-7f62-447b-89d3-134e1ade1471/volumes" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.687021 4715 scope.go:117] "RemoveContainer" containerID="3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18" Dec 03 23:01:39 crc kubenswrapper[4715]: E1203 23:01:39.687532 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18\": container with ID starting with 3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18 not found: ID does not exist" containerID="3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.687565 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18"} err="failed to get container status \"3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18\": rpc error: code = NotFound desc = could not find container \"3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18\": container with ID starting with 3ba44d743e95658463e330d9554b8bcd6fc87dfc1ccb9410732c6b2e50c53d18 not found: ID does not exist" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.687585 4715 scope.go:117] "RemoveContainer" containerID="dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7" Dec 03 23:01:39 crc kubenswrapper[4715]: E1203 23:01:39.687959 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7\": container with ID starting with dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7 not found: ID does not exist" containerID="dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.687982 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7"} err="failed to get container status \"dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7\": rpc error: code = NotFound desc = could not find container \"dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7\": container with ID starting with dc83b6094b85b17afbdf61f5f88e594e0bd094599c03d1a1c745bd58e8a30cf7 not found: ID does not exist" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.687997 4715 scope.go:117] "RemoveContainer" containerID="07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877" Dec 03 23:01:39 crc kubenswrapper[4715]: E1203 23:01:39.688236 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877\": container with ID starting with 07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877 not found: ID does not exist" containerID="07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877" Dec 03 23:01:39 crc kubenswrapper[4715]: I1203 23:01:39.688256 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877"} err="failed to get container status \"07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877\": rpc error: code = NotFound desc = could not find container \"07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877\": container with ID starting with 07d8db9887ad4656ad9a2d50adc033c66bc40818efa7f5f27b641fa4f7433877 not found: ID does not exist" Dec 03 23:01:44 crc kubenswrapper[4715]: E1203 23:01:44.601059 4715 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.50:48676->38.102.83.50:43647: read tcp 38.102.83.50:48676->38.102.83.50:43647: read: connection reset by peer Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.160533 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.161148 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.161210 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.162189 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0ba545ca47db831472f2241785c844e2b365936c2d043769501a6e2a2e23232d"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.162247 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://0ba545ca47db831472f2241785c844e2b365936c2d043769501a6e2a2e23232d" gracePeriod=600 Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.812911 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="0ba545ca47db831472f2241785c844e2b365936c2d043769501a6e2a2e23232d" exitCode=0 Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.812964 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"0ba545ca47db831472f2241785c844e2b365936c2d043769501a6e2a2e23232d"} Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.813684 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b"} Dec 03 23:02:05 crc kubenswrapper[4715]: I1203 23:02:05.813744 4715 scope.go:117] "RemoveContainer" containerID="8601bd61f28ec1d5795e66870c8e04fd8626aae32f047c1202a337667b36b828" Dec 03 23:02:58 crc kubenswrapper[4715]: I1203 23:02:58.302386 4715 generic.go:334] "Generic (PLEG): container finished" podID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerID="4ecbfed57d7dd4008ef3432170d724b5f180638eb6db310a58d80b788dd0d852" exitCode=0 Dec 03 23:02:58 crc kubenswrapper[4715]: I1203 23:02:58.302477 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-kj54w/must-gather-v5xct" event={"ID":"d0ca9866-708a-48c1-847b-bc7cb8000b02","Type":"ContainerDied","Data":"4ecbfed57d7dd4008ef3432170d724b5f180638eb6db310a58d80b788dd0d852"} Dec 03 23:02:58 crc kubenswrapper[4715]: I1203 23:02:58.303760 4715 scope.go:117] "RemoveContainer" containerID="4ecbfed57d7dd4008ef3432170d724b5f180638eb6db310a58d80b788dd0d852" Dec 03 23:02:59 crc kubenswrapper[4715]: I1203 23:02:59.069064 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kj54w_must-gather-v5xct_d0ca9866-708a-48c1-847b-bc7cb8000b02/gather/0.log" Dec 03 23:03:07 crc kubenswrapper[4715]: I1203 23:03:07.702901 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-kj54w/must-gather-v5xct"] Dec 03 23:03:07 crc kubenswrapper[4715]: I1203 23:03:07.703811 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-kj54w/must-gather-v5xct" podUID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerName="copy" containerID="cri-o://e6f96deada2c7c60fab4e6931faf7060d7b80dac52e23bdf2286820864a70e8d" gracePeriod=2 Dec 03 23:03:07 crc kubenswrapper[4715]: I1203 23:03:07.713819 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-kj54w/must-gather-v5xct"] Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.448423 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kj54w_must-gather-v5xct_d0ca9866-708a-48c1-847b-bc7cb8000b02/copy/0.log" Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.449572 4715 generic.go:334] "Generic (PLEG): container finished" podID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerID="e6f96deada2c7c60fab4e6931faf7060d7b80dac52e23bdf2286820864a70e8d" exitCode=143 Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.622899 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kj54w_must-gather-v5xct_d0ca9866-708a-48c1-847b-bc7cb8000b02/copy/0.log" Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.623527 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.762971 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmksc\" (UniqueName: \"kubernetes.io/projected/d0ca9866-708a-48c1-847b-bc7cb8000b02-kube-api-access-wmksc\") pod \"d0ca9866-708a-48c1-847b-bc7cb8000b02\" (UID: \"d0ca9866-708a-48c1-847b-bc7cb8000b02\") " Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.763146 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d0ca9866-708a-48c1-847b-bc7cb8000b02-must-gather-output\") pod \"d0ca9866-708a-48c1-847b-bc7cb8000b02\" (UID: \"d0ca9866-708a-48c1-847b-bc7cb8000b02\") " Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.768816 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0ca9866-708a-48c1-847b-bc7cb8000b02-kube-api-access-wmksc" (OuterVolumeSpecName: "kube-api-access-wmksc") pod "d0ca9866-708a-48c1-847b-bc7cb8000b02" (UID: "d0ca9866-708a-48c1-847b-bc7cb8000b02"). InnerVolumeSpecName "kube-api-access-wmksc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.865696 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmksc\" (UniqueName: \"kubernetes.io/projected/d0ca9866-708a-48c1-847b-bc7cb8000b02-kube-api-access-wmksc\") on node \"crc\" DevicePath \"\"" Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.887370 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0ca9866-708a-48c1-847b-bc7cb8000b02-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "d0ca9866-708a-48c1-847b-bc7cb8000b02" (UID: "d0ca9866-708a-48c1-847b-bc7cb8000b02"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:03:08 crc kubenswrapper[4715]: I1203 23:03:08.967769 4715 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/d0ca9866-708a-48c1-847b-bc7cb8000b02-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 23:03:09 crc kubenswrapper[4715]: I1203 23:03:09.460935 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-kj54w_must-gather-v5xct_d0ca9866-708a-48c1-847b-bc7cb8000b02/copy/0.log" Dec 03 23:03:09 crc kubenswrapper[4715]: I1203 23:03:09.461761 4715 scope.go:117] "RemoveContainer" containerID="e6f96deada2c7c60fab4e6931faf7060d7b80dac52e23bdf2286820864a70e8d" Dec 03 23:03:09 crc kubenswrapper[4715]: I1203 23:03:09.461818 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-kj54w/must-gather-v5xct" Dec 03 23:03:09 crc kubenswrapper[4715]: I1203 23:03:09.483178 4715 scope.go:117] "RemoveContainer" containerID="4ecbfed57d7dd4008ef3432170d724b5f180638eb6db310a58d80b788dd0d852" Dec 03 23:03:09 crc kubenswrapper[4715]: I1203 23:03:09.646856 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0ca9866-708a-48c1-847b-bc7cb8000b02" path="/var/lib/kubelet/pods/d0ca9866-708a-48c1-847b-bc7cb8000b02/volumes" Dec 03 23:04:05 crc kubenswrapper[4715]: I1203 23:04:05.160078 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:04:05 crc kubenswrapper[4715]: I1203 23:04:05.160705 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.839188 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fp4zk"] Dec 03 23:04:10 crc kubenswrapper[4715]: E1203 23:04:10.840098 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="011252d1-7f62-447b-89d3-134e1ade1471" containerName="extract-content" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.840112 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="011252d1-7f62-447b-89d3-134e1ade1471" containerName="extract-content" Dec 03 23:04:10 crc kubenswrapper[4715]: E1203 23:04:10.840139 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerName="copy" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.840145 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerName="copy" Dec 03 23:04:10 crc kubenswrapper[4715]: E1203 23:04:10.840160 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerName="gather" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.840167 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerName="gather" Dec 03 23:04:10 crc kubenswrapper[4715]: E1203 23:04:10.840176 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="011252d1-7f62-447b-89d3-134e1ade1471" containerName="extract-utilities" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.840181 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="011252d1-7f62-447b-89d3-134e1ade1471" containerName="extract-utilities" Dec 03 23:04:10 crc kubenswrapper[4715]: E1203 23:04:10.840196 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="011252d1-7f62-447b-89d3-134e1ade1471" containerName="registry-server" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.840202 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="011252d1-7f62-447b-89d3-134e1ade1471" containerName="registry-server" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.840378 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerName="copy" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.840396 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="011252d1-7f62-447b-89d3-134e1ade1471" containerName="registry-server" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.840408 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0ca9866-708a-48c1-847b-bc7cb8000b02" containerName="gather" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.842123 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.856974 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fp4zk"] Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.903496 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-utilities\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.903687 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgwvg\" (UniqueName: \"kubernetes.io/projected/0e47cb33-e97e-4222-a162-f0191ce3874f-kube-api-access-xgwvg\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:10 crc kubenswrapper[4715]: I1203 23:04:10.903888 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-catalog-content\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:11 crc kubenswrapper[4715]: I1203 23:04:11.006412 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-catalog-content\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:11 crc kubenswrapper[4715]: I1203 23:04:11.006563 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-utilities\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:11 crc kubenswrapper[4715]: I1203 23:04:11.006683 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgwvg\" (UniqueName: \"kubernetes.io/projected/0e47cb33-e97e-4222-a162-f0191ce3874f-kube-api-access-xgwvg\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:11 crc kubenswrapper[4715]: I1203 23:04:11.007039 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-catalog-content\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:11 crc kubenswrapper[4715]: I1203 23:04:11.007215 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-utilities\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:11 crc kubenswrapper[4715]: I1203 23:04:11.032598 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgwvg\" (UniqueName: \"kubernetes.io/projected/0e47cb33-e97e-4222-a162-f0191ce3874f-kube-api-access-xgwvg\") pod \"redhat-operators-fp4zk\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:11 crc kubenswrapper[4715]: I1203 23:04:11.165346 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:11 crc kubenswrapper[4715]: I1203 23:04:11.607201 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fp4zk"] Dec 03 23:04:13 crc kubenswrapper[4715]: I1203 23:04:13.037953 4715 generic.go:334] "Generic (PLEG): container finished" podID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerID="6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba" exitCode=0 Dec 03 23:04:13 crc kubenswrapper[4715]: I1203 23:04:13.038052 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp4zk" event={"ID":"0e47cb33-e97e-4222-a162-f0191ce3874f","Type":"ContainerDied","Data":"6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba"} Dec 03 23:04:13 crc kubenswrapper[4715]: I1203 23:04:13.038535 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp4zk" event={"ID":"0e47cb33-e97e-4222-a162-f0191ce3874f","Type":"ContainerStarted","Data":"e2db2b6c34a15cc1e9cfb8f84097e368ef9e5500a18c5baaff216eb9563250c3"} Dec 03 23:04:13 crc kubenswrapper[4715]: I1203 23:04:13.040220 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 23:04:14 crc kubenswrapper[4715]: I1203 23:04:14.047383 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp4zk" event={"ID":"0e47cb33-e97e-4222-a162-f0191ce3874f","Type":"ContainerStarted","Data":"ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e"} Dec 03 23:04:15 crc kubenswrapper[4715]: I1203 23:04:15.057112 4715 generic.go:334] "Generic (PLEG): container finished" podID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerID="ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e" exitCode=0 Dec 03 23:04:15 crc kubenswrapper[4715]: I1203 23:04:15.057220 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp4zk" event={"ID":"0e47cb33-e97e-4222-a162-f0191ce3874f","Type":"ContainerDied","Data":"ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e"} Dec 03 23:04:16 crc kubenswrapper[4715]: I1203 23:04:16.068548 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp4zk" event={"ID":"0e47cb33-e97e-4222-a162-f0191ce3874f","Type":"ContainerStarted","Data":"fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f"} Dec 03 23:04:16 crc kubenswrapper[4715]: I1203 23:04:16.096786 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fp4zk" podStartSLOduration=3.557689848 podStartE2EDuration="6.096765395s" podCreationTimestamp="2025-12-03 23:04:10 +0000 UTC" firstStartedPulling="2025-12-03 23:04:13.039970961 +0000 UTC m=+4949.782681556" lastFinishedPulling="2025-12-03 23:04:15.579046508 +0000 UTC m=+4952.321757103" observedRunningTime="2025-12-03 23:04:16.087726369 +0000 UTC m=+4952.830436964" watchObservedRunningTime="2025-12-03 23:04:16.096765395 +0000 UTC m=+4952.839475990" Dec 03 23:04:21 crc kubenswrapper[4715]: I1203 23:04:21.166433 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:21 crc kubenswrapper[4715]: I1203 23:04:21.166929 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:21 crc kubenswrapper[4715]: I1203 23:04:21.220567 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:22 crc kubenswrapper[4715]: I1203 23:04:22.172388 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:22 crc kubenswrapper[4715]: I1203 23:04:22.228576 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fp4zk"] Dec 03 23:04:24 crc kubenswrapper[4715]: I1203 23:04:24.131728 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fp4zk" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerName="registry-server" containerID="cri-o://fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f" gracePeriod=2 Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.017045 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.143324 4715 generic.go:334] "Generic (PLEG): container finished" podID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerID="fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f" exitCode=0 Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.143410 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fp4zk" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.143438 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp4zk" event={"ID":"0e47cb33-e97e-4222-a162-f0191ce3874f","Type":"ContainerDied","Data":"fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f"} Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.143933 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fp4zk" event={"ID":"0e47cb33-e97e-4222-a162-f0191ce3874f","Type":"ContainerDied","Data":"e2db2b6c34a15cc1e9cfb8f84097e368ef9e5500a18c5baaff216eb9563250c3"} Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.143972 4715 scope.go:117] "RemoveContainer" containerID="fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.164097 4715 scope.go:117] "RemoveContainer" containerID="ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.170156 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xgwvg\" (UniqueName: \"kubernetes.io/projected/0e47cb33-e97e-4222-a162-f0191ce3874f-kube-api-access-xgwvg\") pod \"0e47cb33-e97e-4222-a162-f0191ce3874f\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.170433 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-utilities\") pod \"0e47cb33-e97e-4222-a162-f0191ce3874f\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.170455 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-catalog-content\") pod \"0e47cb33-e97e-4222-a162-f0191ce3874f\" (UID: \"0e47cb33-e97e-4222-a162-f0191ce3874f\") " Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.171349 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-utilities" (OuterVolumeSpecName: "utilities") pod "0e47cb33-e97e-4222-a162-f0191ce3874f" (UID: "0e47cb33-e97e-4222-a162-f0191ce3874f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.177211 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e47cb33-e97e-4222-a162-f0191ce3874f-kube-api-access-xgwvg" (OuterVolumeSpecName: "kube-api-access-xgwvg") pod "0e47cb33-e97e-4222-a162-f0191ce3874f" (UID: "0e47cb33-e97e-4222-a162-f0191ce3874f"). InnerVolumeSpecName "kube-api-access-xgwvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.186445 4715 scope.go:117] "RemoveContainer" containerID="6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.269955 4715 scope.go:117] "RemoveContainer" containerID="fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f" Dec 03 23:04:25 crc kubenswrapper[4715]: E1203 23:04:25.270367 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f\": container with ID starting with fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f not found: ID does not exist" containerID="fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.270407 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f"} err="failed to get container status \"fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f\": rpc error: code = NotFound desc = could not find container \"fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f\": container with ID starting with fcd7dba216ab6db8beead1a02f26c18f719748b9146be17bda6cf6fda0827b1f not found: ID does not exist" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.270434 4715 scope.go:117] "RemoveContainer" containerID="ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e" Dec 03 23:04:25 crc kubenswrapper[4715]: E1203 23:04:25.270708 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e\": container with ID starting with ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e not found: ID does not exist" containerID="ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.270729 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e"} err="failed to get container status \"ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e\": rpc error: code = NotFound desc = could not find container \"ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e\": container with ID starting with ce0b9055ca02ddb3c22a8a3e0a9610a476030a4a526e9c6330300ccf67ae6b6e not found: ID does not exist" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.270746 4715 scope.go:117] "RemoveContainer" containerID="6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba" Dec 03 23:04:25 crc kubenswrapper[4715]: E1203 23:04:25.271007 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba\": container with ID starting with 6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba not found: ID does not exist" containerID="6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.271042 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba"} err="failed to get container status \"6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba\": rpc error: code = NotFound desc = could not find container \"6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba\": container with ID starting with 6370c26ac886af56bf59b5d7c53d658954982624d612d6f5e583b2beca5e4bba not found: ID does not exist" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.272196 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.272230 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xgwvg\" (UniqueName: \"kubernetes.io/projected/0e47cb33-e97e-4222-a162-f0191ce3874f-kube-api-access-xgwvg\") on node \"crc\" DevicePath \"\"" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.399360 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e47cb33-e97e-4222-a162-f0191ce3874f" (UID: "0e47cb33-e97e-4222-a162-f0191ce3874f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.476075 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e47cb33-e97e-4222-a162-f0191ce3874f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.476911 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fp4zk"] Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.487970 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fp4zk"] Dec 03 23:04:25 crc kubenswrapper[4715]: I1203 23:04:25.646953 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" path="/var/lib/kubelet/pods/0e47cb33-e97e-4222-a162-f0191ce3874f/volumes" Dec 03 23:04:35 crc kubenswrapper[4715]: I1203 23:04:35.159699 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:04:35 crc kubenswrapper[4715]: I1203 23:04:35.160251 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:04:56 crc kubenswrapper[4715]: I1203 23:04:56.818125 4715 scope.go:117] "RemoveContainer" containerID="88b53ac05184d80eb238e26e17b9d0d6587ca98597bb83ae276dde0c50e9fc90" Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.159737 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.160438 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.160542 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.161490 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.161592 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" gracePeriod=600 Dec 03 23:05:05 crc kubenswrapper[4715]: E1203 23:05:05.293568 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.500604 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" exitCode=0 Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.500645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b"} Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.500676 4715 scope.go:117] "RemoveContainer" containerID="0ba545ca47db831472f2241785c844e2b365936c2d043769501a6e2a2e23232d" Dec 03 23:05:05 crc kubenswrapper[4715]: I1203 23:05:05.501300 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:05:05 crc kubenswrapper[4715]: E1203 23:05:05.501626 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:05:20 crc kubenswrapper[4715]: I1203 23:05:20.635156 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:05:20 crc kubenswrapper[4715]: E1203 23:05:20.636542 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:05:33 crc kubenswrapper[4715]: I1203 23:05:33.642261 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:05:33 crc kubenswrapper[4715]: E1203 23:05:33.643016 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:05:46 crc kubenswrapper[4715]: I1203 23:05:46.633983 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:05:46 crc kubenswrapper[4715]: E1203 23:05:46.634673 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:05:56 crc kubenswrapper[4715]: I1203 23:05:56.889427 4715 scope.go:117] "RemoveContainer" containerID="077df71e4964695572bfa2c51daf38b5907c35d62082bb08eb6d3081a352ed2b" Dec 03 23:05:56 crc kubenswrapper[4715]: I1203 23:05:56.915238 4715 scope.go:117] "RemoveContainer" containerID="c1642f36dd288afa37f27a2367b9230b08ce5b1e332c93ad4648b57e15bb7971" Dec 03 23:05:56 crc kubenswrapper[4715]: I1203 23:05:56.965478 4715 scope.go:117] "RemoveContainer" containerID="e931501bb7009123cd5935f42780627813fae05c12a463442102e8679aefc734" Dec 03 23:05:58 crc kubenswrapper[4715]: I1203 23:05:58.634479 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:05:58 crc kubenswrapper[4715]: E1203 23:05:58.634760 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:06:11 crc kubenswrapper[4715]: I1203 23:06:11.635073 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:06:11 crc kubenswrapper[4715]: E1203 23:06:11.636333 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:06:25 crc kubenswrapper[4715]: I1203 23:06:25.634630 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:06:25 crc kubenswrapper[4715]: E1203 23:06:25.635465 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:06:37 crc kubenswrapper[4715]: I1203 23:06:37.633923 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:06:37 crc kubenswrapper[4715]: E1203 23:06:37.634748 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:06:49 crc kubenswrapper[4715]: I1203 23:06:49.645138 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:06:49 crc kubenswrapper[4715]: E1203 23:06:49.646017 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:07:01 crc kubenswrapper[4715]: I1203 23:07:01.634892 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:07:01 crc kubenswrapper[4715]: E1203 23:07:01.635950 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:07:13 crc kubenswrapper[4715]: I1203 23:07:13.641863 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:07:13 crc kubenswrapper[4715]: E1203 23:07:13.642804 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.243301 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-btdhv/must-gather-pc8sw"] Dec 03 23:07:25 crc kubenswrapper[4715]: E1203 23:07:25.244340 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerName="registry-server" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.244360 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerName="registry-server" Dec 03 23:07:25 crc kubenswrapper[4715]: E1203 23:07:25.244391 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerName="extract-content" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.244399 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerName="extract-content" Dec 03 23:07:25 crc kubenswrapper[4715]: E1203 23:07:25.244421 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerName="extract-utilities" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.244430 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerName="extract-utilities" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.244690 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e47cb33-e97e-4222-a162-f0191ce3874f" containerName="registry-server" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.246005 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.251328 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-btdhv/must-gather-pc8sw"] Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.254014 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-btdhv"/"kube-root-ca.crt" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.254113 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-btdhv"/"openshift-service-ca.crt" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.380094 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k9w2\" (UniqueName: \"kubernetes.io/projected/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-kube-api-access-2k9w2\") pod \"must-gather-pc8sw\" (UID: \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\") " pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.380158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-must-gather-output\") pod \"must-gather-pc8sw\" (UID: \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\") " pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.482281 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k9w2\" (UniqueName: \"kubernetes.io/projected/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-kube-api-access-2k9w2\") pod \"must-gather-pc8sw\" (UID: \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\") " pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.482361 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-must-gather-output\") pod \"must-gather-pc8sw\" (UID: \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\") " pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.482767 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-must-gather-output\") pod \"must-gather-pc8sw\" (UID: \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\") " pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.499726 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k9w2\" (UniqueName: \"kubernetes.io/projected/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-kube-api-access-2k9w2\") pod \"must-gather-pc8sw\" (UID: \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\") " pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:07:25 crc kubenswrapper[4715]: I1203 23:07:25.585335 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:07:26 crc kubenswrapper[4715]: I1203 23:07:26.022995 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-btdhv/must-gather-pc8sw"] Dec 03 23:07:27 crc kubenswrapper[4715]: I1203 23:07:27.007306 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/must-gather-pc8sw" event={"ID":"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b","Type":"ContainerStarted","Data":"33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1"} Dec 03 23:07:27 crc kubenswrapper[4715]: I1203 23:07:27.007368 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/must-gather-pc8sw" event={"ID":"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b","Type":"ContainerStarted","Data":"d0f41ceeac97cea60454107b15f04f16e7842986d0566f2d1674c41adad34082"} Dec 03 23:07:27 crc kubenswrapper[4715]: I1203 23:07:27.634723 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:07:27 crc kubenswrapper[4715]: E1203 23:07:27.634978 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:07:28 crc kubenswrapper[4715]: I1203 23:07:28.019527 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/must-gather-pc8sw" event={"ID":"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b","Type":"ContainerStarted","Data":"393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127"} Dec 03 23:07:28 crc kubenswrapper[4715]: I1203 23:07:28.040182 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-btdhv/must-gather-pc8sw" podStartSLOduration=3.040160634 podStartE2EDuration="3.040160634s" podCreationTimestamp="2025-12-03 23:07:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 23:07:28.030929115 +0000 UTC m=+5144.773639710" watchObservedRunningTime="2025-12-03 23:07:28.040160634 +0000 UTC m=+5144.782871229" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.087017 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-btdhv/crc-debug-pw7hd"] Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.088817 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.090940 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-btdhv"/"default-dockercfg-8fjf5" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.165450 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2r8k\" (UniqueName: \"kubernetes.io/projected/cc7368e9-de34-4f4b-85ec-8f03540c87cc-kube-api-access-w2r8k\") pod \"crc-debug-pw7hd\" (UID: \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\") " pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.165824 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc7368e9-de34-4f4b-85ec-8f03540c87cc-host\") pod \"crc-debug-pw7hd\" (UID: \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\") " pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.267489 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2r8k\" (UniqueName: \"kubernetes.io/projected/cc7368e9-de34-4f4b-85ec-8f03540c87cc-kube-api-access-w2r8k\") pod \"crc-debug-pw7hd\" (UID: \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\") " pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.267631 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc7368e9-de34-4f4b-85ec-8f03540c87cc-host\") pod \"crc-debug-pw7hd\" (UID: \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\") " pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.267816 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc7368e9-de34-4f4b-85ec-8f03540c87cc-host\") pod \"crc-debug-pw7hd\" (UID: \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\") " pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.290690 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2r8k\" (UniqueName: \"kubernetes.io/projected/cc7368e9-de34-4f4b-85ec-8f03540c87cc-kube-api-access-w2r8k\") pod \"crc-debug-pw7hd\" (UID: \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\") " pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:07:30 crc kubenswrapper[4715]: I1203 23:07:30.410236 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:07:30 crc kubenswrapper[4715]: W1203 23:07:30.451343 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc7368e9_de34_4f4b_85ec_8f03540c87cc.slice/crio-e243c667881b75ee93ee441b3aef6c6fb3807aedbdeb16064020be45cbd6d5b8 WatchSource:0}: Error finding container e243c667881b75ee93ee441b3aef6c6fb3807aedbdeb16064020be45cbd6d5b8: Status 404 returned error can't find the container with id e243c667881b75ee93ee441b3aef6c6fb3807aedbdeb16064020be45cbd6d5b8 Dec 03 23:07:31 crc kubenswrapper[4715]: I1203 23:07:31.045017 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/crc-debug-pw7hd" event={"ID":"cc7368e9-de34-4f4b-85ec-8f03540c87cc","Type":"ContainerStarted","Data":"95d857137edda94c308409535df37e7704d5cd1435136a1e7f9ce3df6511422d"} Dec 03 23:07:31 crc kubenswrapper[4715]: I1203 23:07:31.045396 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/crc-debug-pw7hd" event={"ID":"cc7368e9-de34-4f4b-85ec-8f03540c87cc","Type":"ContainerStarted","Data":"e243c667881b75ee93ee441b3aef6c6fb3807aedbdeb16064020be45cbd6d5b8"} Dec 03 23:07:31 crc kubenswrapper[4715]: I1203 23:07:31.058594 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-btdhv/crc-debug-pw7hd" podStartSLOduration=1.058577028 podStartE2EDuration="1.058577028s" podCreationTimestamp="2025-12-03 23:07:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 23:07:31.057802458 +0000 UTC m=+5147.800513053" watchObservedRunningTime="2025-12-03 23:07:31.058577028 +0000 UTC m=+5147.801287623" Dec 03 23:07:39 crc kubenswrapper[4715]: I1203 23:07:39.636861 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:07:39 crc kubenswrapper[4715]: E1203 23:07:39.637521 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:07:54 crc kubenswrapper[4715]: I1203 23:07:54.634801 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:07:54 crc kubenswrapper[4715]: E1203 23:07:54.635565 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:08:05 crc kubenswrapper[4715]: I1203 23:08:05.331266 4715 generic.go:334] "Generic (PLEG): container finished" podID="cc7368e9-de34-4f4b-85ec-8f03540c87cc" containerID="95d857137edda94c308409535df37e7704d5cd1435136a1e7f9ce3df6511422d" exitCode=0 Dec 03 23:08:05 crc kubenswrapper[4715]: I1203 23:08:05.331329 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/crc-debug-pw7hd" event={"ID":"cc7368e9-de34-4f4b-85ec-8f03540c87cc","Type":"ContainerDied","Data":"95d857137edda94c308409535df37e7704d5cd1435136a1e7f9ce3df6511422d"} Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.460477 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.498018 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-btdhv/crc-debug-pw7hd"] Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.507137 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-btdhv/crc-debug-pw7hd"] Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.576000 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc7368e9-de34-4f4b-85ec-8f03540c87cc-host\") pod \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\" (UID: \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\") " Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.576118 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2r8k\" (UniqueName: \"kubernetes.io/projected/cc7368e9-de34-4f4b-85ec-8f03540c87cc-kube-api-access-w2r8k\") pod \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\" (UID: \"cc7368e9-de34-4f4b-85ec-8f03540c87cc\") " Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.576833 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cc7368e9-de34-4f4b-85ec-8f03540c87cc-host" (OuterVolumeSpecName: "host") pod "cc7368e9-de34-4f4b-85ec-8f03540c87cc" (UID: "cc7368e9-de34-4f4b-85ec-8f03540c87cc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.600351 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc7368e9-de34-4f4b-85ec-8f03540c87cc-kube-api-access-w2r8k" (OuterVolumeSpecName: "kube-api-access-w2r8k") pod "cc7368e9-de34-4f4b-85ec-8f03540c87cc" (UID: "cc7368e9-de34-4f4b-85ec-8f03540c87cc"). InnerVolumeSpecName "kube-api-access-w2r8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.679216 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/cc7368e9-de34-4f4b-85ec-8f03540c87cc-host\") on node \"crc\" DevicePath \"\"" Dec 03 23:08:06 crc kubenswrapper[4715]: I1203 23:08:06.679251 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2r8k\" (UniqueName: \"kubernetes.io/projected/cc7368e9-de34-4f4b-85ec-8f03540c87cc-kube-api-access-w2r8k\") on node \"crc\" DevicePath \"\"" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.360355 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e243c667881b75ee93ee441b3aef6c6fb3807aedbdeb16064020be45cbd6d5b8" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.360661 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-pw7hd" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.634859 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:08:07 crc kubenswrapper[4715]: E1203 23:08:07.635408 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.647662 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc7368e9-de34-4f4b-85ec-8f03540c87cc" path="/var/lib/kubelet/pods/cc7368e9-de34-4f4b-85ec-8f03540c87cc/volumes" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.710999 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-btdhv/crc-debug-r6ct2"] Dec 03 23:08:07 crc kubenswrapper[4715]: E1203 23:08:07.711471 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc7368e9-de34-4f4b-85ec-8f03540c87cc" containerName="container-00" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.711515 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc7368e9-de34-4f4b-85ec-8f03540c87cc" containerName="container-00" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.711781 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc7368e9-de34-4f4b-85ec-8f03540c87cc" containerName="container-00" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.712789 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.714808 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-btdhv"/"default-dockercfg-8fjf5" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.899585 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15b10b44-e59f-4158-b062-e2df1d67019c-host\") pod \"crc-debug-r6ct2\" (UID: \"15b10b44-e59f-4158-b062-e2df1d67019c\") " pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:07 crc kubenswrapper[4715]: I1203 23:08:07.899672 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbcvf\" (UniqueName: \"kubernetes.io/projected/15b10b44-e59f-4158-b062-e2df1d67019c-kube-api-access-jbcvf\") pod \"crc-debug-r6ct2\" (UID: \"15b10b44-e59f-4158-b062-e2df1d67019c\") " pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.001374 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbcvf\" (UniqueName: \"kubernetes.io/projected/15b10b44-e59f-4158-b062-e2df1d67019c-kube-api-access-jbcvf\") pod \"crc-debug-r6ct2\" (UID: \"15b10b44-e59f-4158-b062-e2df1d67019c\") " pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.001571 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15b10b44-e59f-4158-b062-e2df1d67019c-host\") pod \"crc-debug-r6ct2\" (UID: \"15b10b44-e59f-4158-b062-e2df1d67019c\") " pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.001756 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15b10b44-e59f-4158-b062-e2df1d67019c-host\") pod \"crc-debug-r6ct2\" (UID: \"15b10b44-e59f-4158-b062-e2df1d67019c\") " pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.025377 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbcvf\" (UniqueName: \"kubernetes.io/projected/15b10b44-e59f-4158-b062-e2df1d67019c-kube-api-access-jbcvf\") pod \"crc-debug-r6ct2\" (UID: \"15b10b44-e59f-4158-b062-e2df1d67019c\") " pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.030806 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.371434 4715 generic.go:334] "Generic (PLEG): container finished" podID="15b10b44-e59f-4158-b062-e2df1d67019c" containerID="716b1f4113002e25dcee8cc5d851e78fb5a2e80a1fcbd022efac865b75165b74" exitCode=0 Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.371520 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/crc-debug-r6ct2" event={"ID":"15b10b44-e59f-4158-b062-e2df1d67019c","Type":"ContainerDied","Data":"716b1f4113002e25dcee8cc5d851e78fb5a2e80a1fcbd022efac865b75165b74"} Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.371852 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/crc-debug-r6ct2" event={"ID":"15b10b44-e59f-4158-b062-e2df1d67019c","Type":"ContainerStarted","Data":"4bf311a580374e317c018c4d322352bd513ee3e5504f2731782d64de02b40da4"} Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.807839 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-btdhv/crc-debug-r6ct2"] Dec 03 23:08:08 crc kubenswrapper[4715]: I1203 23:08:08.815075 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-btdhv/crc-debug-r6ct2"] Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.479944 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.629206 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15b10b44-e59f-4158-b062-e2df1d67019c-host\") pod \"15b10b44-e59f-4158-b062-e2df1d67019c\" (UID: \"15b10b44-e59f-4158-b062-e2df1d67019c\") " Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.629291 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbcvf\" (UniqueName: \"kubernetes.io/projected/15b10b44-e59f-4158-b062-e2df1d67019c-kube-api-access-jbcvf\") pod \"15b10b44-e59f-4158-b062-e2df1d67019c\" (UID: \"15b10b44-e59f-4158-b062-e2df1d67019c\") " Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.629347 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15b10b44-e59f-4158-b062-e2df1d67019c-host" (OuterVolumeSpecName: "host") pod "15b10b44-e59f-4158-b062-e2df1d67019c" (UID: "15b10b44-e59f-4158-b062-e2df1d67019c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.630114 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/15b10b44-e59f-4158-b062-e2df1d67019c-host\") on node \"crc\" DevicePath \"\"" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.638771 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15b10b44-e59f-4158-b062-e2df1d67019c-kube-api-access-jbcvf" (OuterVolumeSpecName: "kube-api-access-jbcvf") pod "15b10b44-e59f-4158-b062-e2df1d67019c" (UID: "15b10b44-e59f-4158-b062-e2df1d67019c"). InnerVolumeSpecName "kube-api-access-jbcvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.658708 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15b10b44-e59f-4158-b062-e2df1d67019c" path="/var/lib/kubelet/pods/15b10b44-e59f-4158-b062-e2df1d67019c/volumes" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.732004 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbcvf\" (UniqueName: \"kubernetes.io/projected/15b10b44-e59f-4158-b062-e2df1d67019c-kube-api-access-jbcvf\") on node \"crc\" DevicePath \"\"" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.989189 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-btdhv/crc-debug-65mgr"] Dec 03 23:08:09 crc kubenswrapper[4715]: E1203 23:08:09.990025 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15b10b44-e59f-4158-b062-e2df1d67019c" containerName="container-00" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.990041 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="15b10b44-e59f-4158-b062-e2df1d67019c" containerName="container-00" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.990360 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="15b10b44-e59f-4158-b062-e2df1d67019c" containerName="container-00" Dec 03 23:08:09 crc kubenswrapper[4715]: I1203 23:08:09.991181 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.138207 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46f66136-98be-47da-8880-9d8e37f6665a-host\") pod \"crc-debug-65mgr\" (UID: \"46f66136-98be-47da-8880-9d8e37f6665a\") " pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.138286 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sthjb\" (UniqueName: \"kubernetes.io/projected/46f66136-98be-47da-8880-9d8e37f6665a-kube-api-access-sthjb\") pod \"crc-debug-65mgr\" (UID: \"46f66136-98be-47da-8880-9d8e37f6665a\") " pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.240014 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46f66136-98be-47da-8880-9d8e37f6665a-host\") pod \"crc-debug-65mgr\" (UID: \"46f66136-98be-47da-8880-9d8e37f6665a\") " pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.240448 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sthjb\" (UniqueName: \"kubernetes.io/projected/46f66136-98be-47da-8880-9d8e37f6665a-kube-api-access-sthjb\") pod \"crc-debug-65mgr\" (UID: \"46f66136-98be-47da-8880-9d8e37f6665a\") " pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.240167 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46f66136-98be-47da-8880-9d8e37f6665a-host\") pod \"crc-debug-65mgr\" (UID: \"46f66136-98be-47da-8880-9d8e37f6665a\") " pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.257731 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sthjb\" (UniqueName: \"kubernetes.io/projected/46f66136-98be-47da-8880-9d8e37f6665a-kube-api-access-sthjb\") pod \"crc-debug-65mgr\" (UID: \"46f66136-98be-47da-8880-9d8e37f6665a\") " pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.318096 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.390164 4715 scope.go:117] "RemoveContainer" containerID="716b1f4113002e25dcee8cc5d851e78fb5a2e80a1fcbd022efac865b75165b74" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.390182 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-r6ct2" Dec 03 23:08:10 crc kubenswrapper[4715]: I1203 23:08:10.391991 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/crc-debug-65mgr" event={"ID":"46f66136-98be-47da-8880-9d8e37f6665a","Type":"ContainerStarted","Data":"b9c5f5ef36265e33c27b3fb2544311095b8ef6f19d411835478476bcdb41367d"} Dec 03 23:08:11 crc kubenswrapper[4715]: I1203 23:08:11.401618 4715 generic.go:334] "Generic (PLEG): container finished" podID="46f66136-98be-47da-8880-9d8e37f6665a" containerID="2b0a0c3b5d46ed2899f714e79d19b56939a1afc9ce0dff33af7ba72ac774a6ba" exitCode=0 Dec 03 23:08:11 crc kubenswrapper[4715]: I1203 23:08:11.401661 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/crc-debug-65mgr" event={"ID":"46f66136-98be-47da-8880-9d8e37f6665a","Type":"ContainerDied","Data":"2b0a0c3b5d46ed2899f714e79d19b56939a1afc9ce0dff33af7ba72ac774a6ba"} Dec 03 23:08:11 crc kubenswrapper[4715]: I1203 23:08:11.447770 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-btdhv/crc-debug-65mgr"] Dec 03 23:08:11 crc kubenswrapper[4715]: I1203 23:08:11.455474 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-btdhv/crc-debug-65mgr"] Dec 03 23:08:12 crc kubenswrapper[4715]: I1203 23:08:12.525137 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:12 crc kubenswrapper[4715]: I1203 23:08:12.586084 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46f66136-98be-47da-8880-9d8e37f6665a-host\") pod \"46f66136-98be-47da-8880-9d8e37f6665a\" (UID: \"46f66136-98be-47da-8880-9d8e37f6665a\") " Dec 03 23:08:12 crc kubenswrapper[4715]: I1203 23:08:12.586174 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46f66136-98be-47da-8880-9d8e37f6665a-host" (OuterVolumeSpecName: "host") pod "46f66136-98be-47da-8880-9d8e37f6665a" (UID: "46f66136-98be-47da-8880-9d8e37f6665a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 23:08:12 crc kubenswrapper[4715]: I1203 23:08:12.586310 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sthjb\" (UniqueName: \"kubernetes.io/projected/46f66136-98be-47da-8880-9d8e37f6665a-kube-api-access-sthjb\") pod \"46f66136-98be-47da-8880-9d8e37f6665a\" (UID: \"46f66136-98be-47da-8880-9d8e37f6665a\") " Dec 03 23:08:12 crc kubenswrapper[4715]: I1203 23:08:12.586761 4715 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/46f66136-98be-47da-8880-9d8e37f6665a-host\") on node \"crc\" DevicePath \"\"" Dec 03 23:08:12 crc kubenswrapper[4715]: I1203 23:08:12.591823 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f66136-98be-47da-8880-9d8e37f6665a-kube-api-access-sthjb" (OuterVolumeSpecName: "kube-api-access-sthjb") pod "46f66136-98be-47da-8880-9d8e37f6665a" (UID: "46f66136-98be-47da-8880-9d8e37f6665a"). InnerVolumeSpecName "kube-api-access-sthjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:08:12 crc kubenswrapper[4715]: I1203 23:08:12.688939 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sthjb\" (UniqueName: \"kubernetes.io/projected/46f66136-98be-47da-8880-9d8e37f6665a-kube-api-access-sthjb\") on node \"crc\" DevicePath \"\"" Dec 03 23:08:13 crc kubenswrapper[4715]: I1203 23:08:13.420575 4715 scope.go:117] "RemoveContainer" containerID="2b0a0c3b5d46ed2899f714e79d19b56939a1afc9ce0dff33af7ba72ac774a6ba" Dec 03 23:08:13 crc kubenswrapper[4715]: I1203 23:08:13.420731 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/crc-debug-65mgr" Dec 03 23:08:13 crc kubenswrapper[4715]: I1203 23:08:13.649308 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f66136-98be-47da-8880-9d8e37f6665a" path="/var/lib/kubelet/pods/46f66136-98be-47da-8880-9d8e37f6665a/volumes" Dec 03 23:08:18 crc kubenswrapper[4715]: I1203 23:08:18.634730 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:08:18 crc kubenswrapper[4715]: E1203 23:08:18.635318 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:08:31 crc kubenswrapper[4715]: I1203 23:08:31.635245 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:08:31 crc kubenswrapper[4715]: E1203 23:08:31.636217 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.159602 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-dd5bb56d6-jb5n2_9170a301-98ac-4953-ae87-4910bf0a466a/barbican-api/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.330393 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-dd5bb56d6-jb5n2_9170a301-98ac-4953-ae87-4910bf0a466a/barbican-api-log/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.360442 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9cc86b6b-6xlfj_0e740246-a3e9-4692-8888-e12768f0573e/barbican-keystone-listener/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.418012 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-9cc86b6b-6xlfj_0e740246-a3e9-4692-8888-e12768f0573e/barbican-keystone-listener-log/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.512358 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-78449cf9c5-v4vnx_4382c9dd-8a2d-4634-b9de-0796a994d491/barbican-worker/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.549549 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-78449cf9c5-v4vnx_4382c9dd-8a2d-4634-b9de-0796a994d491/barbican-worker-log/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.722301 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-cvfxx_7d5be5a7-46b7-4c15-b6ef-09e00d9a5b63/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.778653 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ac3428ad-2684-4ca1-ab1e-254306dc84a3/ceilometer-central-agent/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.877650 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ac3428ad-2684-4ca1-ab1e-254306dc84a3/ceilometer-notification-agent/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.906909 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ac3428ad-2684-4ca1-ab1e-254306dc84a3/proxy-httpd/0.log" Dec 03 23:08:37 crc kubenswrapper[4715]: I1203 23:08:37.974585 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ac3428ad-2684-4ca1-ab1e-254306dc84a3/sg-core/0.log" Dec 03 23:08:38 crc kubenswrapper[4715]: I1203 23:08:38.117394 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6cfc1c9d-e789-449f-a73d-a4aead3a03f1/cinder-api/0.log" Dec 03 23:08:38 crc kubenswrapper[4715]: I1203 23:08:38.317627 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6cfc1c9d-e789-449f-a73d-a4aead3a03f1/cinder-api-log/0.log" Dec 03 23:08:38 crc kubenswrapper[4715]: I1203 23:08:38.424016 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cad3e724-7c3e-498d-9a7a-362a4f57c2b5/cinder-scheduler/0.log" Dec 03 23:08:38 crc kubenswrapper[4715]: I1203 23:08:38.517634 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_cad3e724-7c3e-498d-9a7a-362a4f57c2b5/probe/0.log" Dec 03 23:08:38 crc kubenswrapper[4715]: I1203 23:08:38.602727 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-mvtmg_7fa5989a-f985-4757-84dd-3c658472a85e/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:38 crc kubenswrapper[4715]: I1203 23:08:38.711116 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-9mszz_b1a79612-747c-4bbd-be8c-f79a9da177db/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:38 crc kubenswrapper[4715]: I1203 23:08:38.802005 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-bhvj6_a1e5e31a-ff62-48c9-8908-15ff570bda0d/init/0.log" Dec 03 23:08:38 crc kubenswrapper[4715]: I1203 23:08:38.971235 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-bhvj6_a1e5e31a-ff62-48c9-8908-15ff570bda0d/init/0.log" Dec 03 23:08:39 crc kubenswrapper[4715]: I1203 23:08:39.014189 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-78c64bc9c5-bhvj6_a1e5e31a-ff62-48c9-8908-15ff570bda0d/dnsmasq-dns/0.log" Dec 03 23:08:39 crc kubenswrapper[4715]: I1203 23:08:39.062548 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-b2hr5_54307606-6028-435b-bb67-3dd55f7a938a/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:39 crc kubenswrapper[4715]: I1203 23:08:39.234942 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_99136b0c-47a6-4503-806e-ec70fa8c2409/glance-log/0.log" Dec 03 23:08:39 crc kubenswrapper[4715]: I1203 23:08:39.418744 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_20ba12ff-26f9-4100-b21d-0545d33fc7d1/glance-httpd/0.log" Dec 03 23:08:39 crc kubenswrapper[4715]: I1203 23:08:39.461837 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_20ba12ff-26f9-4100-b21d-0545d33fc7d1/glance-log/0.log" Dec 03 23:08:39 crc kubenswrapper[4715]: I1203 23:08:39.658754 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_99136b0c-47a6-4503-806e-ec70fa8c2409/glance-httpd/0.log" Dec 03 23:08:39 crc kubenswrapper[4715]: I1203 23:08:39.778553 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-656478594d-5tbrm_6c7cfacf-a283-45d3-982d-b28e62a9146b/horizon/0.log" Dec 03 23:08:39 crc kubenswrapper[4715]: I1203 23:08:39.852299 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-c27sm_c58da0a9-c5be-4018-a322-7ba8ef960d52/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.065879 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-pmd4j_3976028c-4d85-4518-9151-0e7c94638d58/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.206745 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-656478594d-5tbrm_6c7cfacf-a283-45d3-982d-b28e62a9146b/horizon-log/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.303040 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413381-ncq8v_a2c5df74-94f6-4aa7-a66a-386414a454ce/keystone-cron/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.381537 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-76bcbf7d4d-l7qmc_954b69af-5e1b-4eec-a783-1cbe28e7c0ca/keystone-api/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.410747 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_11b5a8ae-b831-453c-964b-1631e5d1280c/kube-state-metrics/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.552690 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-nt4s4_e419d6c2-a6b7-474d-aea8-6f90eb7ce85a/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.872130 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-666d579d9-hnzld_58badebd-4a34-4430-96ee-706deee12007/neutron-httpd/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.891302 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-wjs5d_c3439842-caec-430b-a91e-0b03443eff15/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:40 crc kubenswrapper[4715]: I1203 23:08:40.934649 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-666d579d9-hnzld_58badebd-4a34-4430-96ee-706deee12007/neutron-api/0.log" Dec 03 23:08:41 crc kubenswrapper[4715]: I1203 23:08:41.414598 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b079ad3a-bcab-48cc-99da-f7f656195558/nova-api-log/0.log" Dec 03 23:08:41 crc kubenswrapper[4715]: I1203 23:08:41.541942 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_c328932f-799e-4f62-88c7-34c17ab3c0bc/nova-cell0-conductor-conductor/0.log" Dec 03 23:08:41 crc kubenswrapper[4715]: I1203 23:08:41.715493 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_5b686fe6-ab19-4e75-9328-7ae61b8b609f/nova-cell1-conductor-conductor/0.log" Dec 03 23:08:41 crc kubenswrapper[4715]: I1203 23:08:41.840106 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_b079ad3a-bcab-48cc-99da-f7f656195558/nova-api-api/0.log" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.014263 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-njnmd_10aa3b67-88a4-4a5d-9d81-096c5b944fb7/nova-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.029089 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e4f46a99-13e6-4175-939a-3ffb54503ef2/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.264125 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0a252fdb-1b30-4278-8bfc-471881a27f21/nova-metadata-log/0.log" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.519192 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b487c9d6-b74d-4db6-8ef4-caee02c83106/mysql-bootstrap/0.log" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.566990 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_5e79bc96-17dc-43d2-b049-da6de097be5e/nova-scheduler-scheduler/0.log" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.633730 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:08:42 crc kubenswrapper[4715]: E1203 23:08:42.634021 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.690429 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b487c9d6-b74d-4db6-8ef4-caee02c83106/mysql-bootstrap/0.log" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.719550 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b487c9d6-b74d-4db6-8ef4-caee02c83106/galera/0.log" Dec 03 23:08:42 crc kubenswrapper[4715]: I1203 23:08:42.923491 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b024438-6200-45dd-9297-ae2b071dae15/mysql-bootstrap/0.log" Dec 03 23:08:43 crc kubenswrapper[4715]: I1203 23:08:43.114134 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b024438-6200-45dd-9297-ae2b071dae15/mysql-bootstrap/0.log" Dec 03 23:08:43 crc kubenswrapper[4715]: I1203 23:08:43.174684 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3b024438-6200-45dd-9297-ae2b071dae15/galera/0.log" Dec 03 23:08:43 crc kubenswrapper[4715]: I1203 23:08:43.598488 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_0a252fdb-1b30-4278-8bfc-471881a27f21/nova-metadata-metadata/0.log" Dec 03 23:08:43 crc kubenswrapper[4715]: I1203 23:08:43.707350 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_205269ef-3624-4663-957b-907cf1a7cadc/openstackclient/0.log" Dec 03 23:08:43 crc kubenswrapper[4715]: I1203 23:08:43.768253 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-5fzvj_893e6b97-27e2-4372-8fc6-90562f2781bb/ovn-controller/0.log" Dec 03 23:08:43 crc kubenswrapper[4715]: I1203 23:08:43.910170 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-sgsmm_473308fd-dd98-47fd-8a5e-32ab39bd6730/openstack-network-exporter/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.057173 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xq7sg_d495583b-de78-4dd5-aa7b-70e15ee3cf0e/ovsdb-server-init/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.207092 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xq7sg_d495583b-de78-4dd5-aa7b-70e15ee3cf0e/ovsdb-server-init/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.220429 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xq7sg_d495583b-de78-4dd5-aa7b-70e15ee3cf0e/ovs-vswitchd/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.257450 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xq7sg_d495583b-de78-4dd5-aa7b-70e15ee3cf0e/ovsdb-server/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.470174 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-68qjb_d0c3bcdb-424a-4da7-9917-7a7853067044/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.488933 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_18d99e58-0606-46f0-91ef-553c75d44ba4/openstack-network-exporter/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.514956 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_18d99e58-0606-46f0-91ef-553c75d44ba4/ovn-northd/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.634945 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_af93de35-2858-4a66-80a3-b7be3a28ff3f/openstack-network-exporter/0.log" Dec 03 23:08:44 crc kubenswrapper[4715]: I1203 23:08:44.702774 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_af93de35-2858-4a66-80a3-b7be3a28ff3f/ovsdbserver-nb/0.log" Dec 03 23:08:45 crc kubenswrapper[4715]: I1203 23:08:45.114130 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_695ea890-c971-4f5a-84ca-fbd83fa5d74a/openstack-network-exporter/0.log" Dec 03 23:08:45 crc kubenswrapper[4715]: I1203 23:08:45.126789 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_695ea890-c971-4f5a-84ca-fbd83fa5d74a/ovsdbserver-sb/0.log" Dec 03 23:08:45 crc kubenswrapper[4715]: I1203 23:08:45.265534 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8694d8458b-2x672_661b1086-fc52-475a-a449-c116baa618a2/placement-api/0.log" Dec 03 23:08:45 crc kubenswrapper[4715]: I1203 23:08:45.392306 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-8694d8458b-2x672_661b1086-fc52-475a-a449-c116baa618a2/placement-log/0.log" Dec 03 23:08:45 crc kubenswrapper[4715]: I1203 23:08:45.403355 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cb376763-8ff4-48ce-a374-ed1119c961e9/setup-container/0.log" Dec 03 23:08:45 crc kubenswrapper[4715]: I1203 23:08:45.741849 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_82483123-49b2-496c-96f4-2d9423e2dfeb/setup-container/0.log" Dec 03 23:08:45 crc kubenswrapper[4715]: I1203 23:08:45.760956 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cb376763-8ff4-48ce-a374-ed1119c961e9/rabbitmq/0.log" Dec 03 23:08:45 crc kubenswrapper[4715]: I1203 23:08:45.795479 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cb376763-8ff4-48ce-a374-ed1119c961e9/setup-container/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.017753 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_82483123-49b2-496c-96f4-2d9423e2dfeb/setup-container/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.025091 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7sv74_7790e81c-da01-4013-81ad-32bd29bba24f/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.121305 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_82483123-49b2-496c-96f4-2d9423e2dfeb/rabbitmq/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.192765 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-8d7ms_e1c0be7c-5e1c-43dd-9148-30cd6bc42b93/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.396360 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-jg2l9_4a15fc43-6ab8-4502-a2d3-9c837a74cb90/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.485290 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-9hf4c_849e6369-17eb-4b31-a5cc-f44dc07c6e1a/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.618909 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2jmsp_2e622ea8-2426-4e53-8775-3880cce092b2/ssh-known-hosts-edpm-deployment/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.779264 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5df8bbbf89-2tsr9_1bf3ff4d-5720-4151-ad53-3683f08c21c5/proxy-server/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.943330 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-7f2gr_18eb44f1-da35-4a10-b826-03c94f6b525e/swift-ring-rebalance/0.log" Dec 03 23:08:46 crc kubenswrapper[4715]: I1203 23:08:46.953434 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5df8bbbf89-2tsr9_1bf3ff4d-5720-4151-ad53-3683f08c21c5/proxy-httpd/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.102582 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/account-auditor/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.154533 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/account-reaper/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.226321 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/account-replicator/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.326017 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/container-auditor/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.340576 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/account-server/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.421849 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/container-replicator/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.447144 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/container-server/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.552754 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/container-updater/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.553363 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-auditor/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.666768 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-expirer/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.680604 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-replicator/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.733950 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-server/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.784125 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/object-updater/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.891184 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/rsync/0.log" Dec 03 23:08:47 crc kubenswrapper[4715]: I1203 23:08:47.905579 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_d58c9eac-e46a-4a54-b1eb-27246d9c064e/swift-recon-cron/0.log" Dec 03 23:08:48 crc kubenswrapper[4715]: I1203 23:08:48.024329 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-7gf2x_32684f56-1e44-4d82-b8d2-f1193ba615d2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:48 crc kubenswrapper[4715]: I1203 23:08:48.127577 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_3c5223ae-7341-4aa4-821b-ec5450e37ebc/tempest-tests-tempest-tests-runner/0.log" Dec 03 23:08:48 crc kubenswrapper[4715]: I1203 23:08:48.199446 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_bf66462e-6db6-4370-9071-d4d60593015d/test-operator-logs-container/0.log" Dec 03 23:08:48 crc kubenswrapper[4715]: I1203 23:08:48.366471 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-8v4t2_db7da37a-b417-433c-86e9-a18446b4af27/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 23:08:55 crc kubenswrapper[4715]: I1203 23:08:55.638700 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:08:55 crc kubenswrapper[4715]: E1203 23:08:55.639651 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:09:00 crc kubenswrapper[4715]: I1203 23:09:00.042183 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5ffb2b06-021a-49c0-abfe-6bb5c8acba3d/memcached/0.log" Dec 03 23:09:09 crc kubenswrapper[4715]: I1203 23:09:09.634403 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:09:09 crc kubenswrapper[4715]: E1203 23:09:09.635226 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.157014 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/util/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.320812 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/util/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.338264 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/pull/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.348234 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/pull/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.507321 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/pull/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.530053 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/extract/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.542767 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_464a6078898499d62f261e598549b500e9e20d1f15505973d1dacb2d02xrm88_cc5a2a79-3856-406e-88ac-b6496ea27a64/util/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.685569 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5hzjb_46e49f12-51e7-45ca-a0cb-c7f8a40adf12/kube-rbac-proxy/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.736085 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-mnd86_9c8c4875-3433-4885-9afa-acbfea9d546b/kube-rbac-proxy/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.771639 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5hzjb_46e49f12-51e7-45ca-a0cb-c7f8a40adf12/manager/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.900273 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-sctl9_a91ec3b7-705a-4500-be8e-e8c789a2c973/kube-rbac-proxy/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.901930 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-mnd86_9c8c4875-3433-4885-9afa-acbfea9d546b/manager/0.log" Dec 03 23:09:15 crc kubenswrapper[4715]: I1203 23:09:15.952232 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-sctl9_a91ec3b7-705a-4500-be8e-e8c789a2c973/manager/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.063218 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-d5phr_b8cbfb12-17c2-46bf-baea-ee8ac16db4ee/kube-rbac-proxy/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.185084 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-d5phr_b8cbfb12-17c2-46bf-baea-ee8ac16db4ee/manager/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.268860 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gc2vn_1cba7fbc-0a66-45b4-b20d-33c81d407b96/kube-rbac-proxy/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.286060 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-gc2vn_1cba7fbc-0a66-45b4-b20d-33c81d407b96/manager/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.371545 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-tgkbx_8219810d-511e-4cc8-a018-49f425b339c0/kube-rbac-proxy/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.435185 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-tgkbx_8219810d-511e-4cc8-a018-49f425b339c0/manager/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.596791 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-kwnlp_6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef/kube-rbac-proxy/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.702810 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-9qbjh_95b90127-8325-43c8-977e-8b972b62db2a/kube-rbac-proxy/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.703345 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-kwnlp_6aea4a92-7dfd-4de9-b9a2-f504ed1ac0ef/manager/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.777050 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-9qbjh_95b90127-8325-43c8-977e-8b972b62db2a/manager/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.915720 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qbp4s_305a708b-69fc-4263-a28f-015fe16402f7/kube-rbac-proxy/0.log" Dec 03 23:09:16 crc kubenswrapper[4715]: I1203 23:09:16.937726 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-qbp4s_305a708b-69fc-4263-a28f-015fe16402f7/manager/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.027700 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-9gf6m_9a900ae2-08bc-403b-a613-52e7870eeebf/kube-rbac-proxy/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.078950 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-9gf6m_9a900ae2-08bc-403b-a613-52e7870eeebf/manager/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.142431 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hqgrk_c6d7646e-fb4a-4199-aadb-670000485477/kube-rbac-proxy/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.194617 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-hqgrk_c6d7646e-fb4a-4199-aadb-670000485477/manager/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.302759 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-gv4jj_3091b003-a466-4d34-8a26-d3e68a3639a9/kube-rbac-proxy/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.336218 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-gv4jj_3091b003-a466-4d34-8a26-d3e68a3639a9/manager/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.447396 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-hm2q2_b8304681-abf0-4e5c-aa37-41453f6e7731/kube-rbac-proxy/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.538182 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-hm2q2_b8304681-abf0-4e5c-aa37-41453f6e7731/manager/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.636886 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-9ch2g_594f9dc7-fb50-4241-bb93-90d8039339dc/kube-rbac-proxy/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.675005 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-9ch2g_594f9dc7-fb50-4241-bb93-90d8039339dc/manager/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.736222 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp_dbeb21cb-a705-464d-ace4-90e1ce5084a5/kube-rbac-proxy/0.log" Dec 03 23:09:17 crc kubenswrapper[4715]: I1203 23:09:17.813866 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4rndpp_dbeb21cb-a705-464d-ace4-90e1ce5084a5/manager/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.101063 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-tpfhd_dc2ba455-421c-4b30-9a6c-93945d7d1c9f/registry-server/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.203258 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-85f8f75cbc-4ld8x_d4195e77-3226-4e2f-b29e-b6ddaaa3bef2/operator/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.304173 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-68hlj_c7e152eb-ab3c-427a-a12b-70f46b881eee/kube-rbac-proxy/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.440436 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-68hlj_c7e152eb-ab3c-427a-a12b-70f46b881eee/manager/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.597660 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-hj6qn_6925aa7e-915e-4bbf-a3e1-991ce581b49c/kube-rbac-proxy/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.678828 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-hj6qn_6925aa7e-915e-4bbf-a3e1-991ce581b49c/manager/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.842272 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-hnzzx_1d2cfa37-465d-4b27-8a8b-90038e1f4bf1/operator/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.903450 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-252wh_f621de25-14cd-43f5-a0a8-cf0866337715/kube-rbac-proxy/0.log" Dec 03 23:09:18 crc kubenswrapper[4715]: I1203 23:09:18.982306 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-bd9c944fb-mkhhs_a41a8afb-d557-4caa-bf3e-57a9a5115f57/manager/0.log" Dec 03 23:09:19 crc kubenswrapper[4715]: I1203 23:09:19.013758 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-252wh_f621de25-14cd-43f5-a0a8-cf0866337715/manager/0.log" Dec 03 23:09:19 crc kubenswrapper[4715]: I1203 23:09:19.078523 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-8g9vt_85bafa11-dcf4-4ea4-b68e-e2d4799e813d/kube-rbac-proxy/0.log" Dec 03 23:09:19 crc kubenswrapper[4715]: I1203 23:09:19.142946 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-8g9vt_85bafa11-dcf4-4ea4-b68e-e2d4799e813d/manager/0.log" Dec 03 23:09:19 crc kubenswrapper[4715]: I1203 23:09:19.209810 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-ttfdp_0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e/kube-rbac-proxy/0.log" Dec 03 23:09:19 crc kubenswrapper[4715]: I1203 23:09:19.242210 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-ttfdp_0d4b2d7c-f73c-4ab9-a5e1-775a99673f7e/manager/0.log" Dec 03 23:09:19 crc kubenswrapper[4715]: I1203 23:09:19.342819 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-gtpjt_959f935f-f997-47bc-8798-d54f7b62f2c9/kube-rbac-proxy/0.log" Dec 03 23:09:19 crc kubenswrapper[4715]: I1203 23:09:19.343893 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-gtpjt_959f935f-f997-47bc-8798-d54f7b62f2c9/manager/0.log" Dec 03 23:09:21 crc kubenswrapper[4715]: I1203 23:09:21.638632 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:09:21 crc kubenswrapper[4715]: E1203 23:09:21.639127 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:09:32 crc kubenswrapper[4715]: I1203 23:09:32.635097 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:09:32 crc kubenswrapper[4715]: E1203 23:09:32.635990 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:09:40 crc kubenswrapper[4715]: I1203 23:09:40.516103 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dwjvx_290986b1-4ac0-40df-9294-0cff8d471733/control-plane-machine-set-operator/0.log" Dec 03 23:09:40 crc kubenswrapper[4715]: I1203 23:09:40.672512 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-77n4d_cd7d035c-0bd3-4568-8384-03b463fb0c5c/kube-rbac-proxy/0.log" Dec 03 23:09:40 crc kubenswrapper[4715]: I1203 23:09:40.679905 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-77n4d_cd7d035c-0bd3-4568-8384-03b463fb0c5c/machine-api-operator/0.log" Dec 03 23:09:44 crc kubenswrapper[4715]: I1203 23:09:44.634924 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:09:44 crc kubenswrapper[4715]: E1203 23:09:44.635891 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:09:55 crc kubenswrapper[4715]: I1203 23:09:55.251622 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-rnccs_bec62745-cf3a-4bc6-9fc6-34c8a56301ce/cert-manager-cainjector/0.log" Dec 03 23:09:55 crc kubenswrapper[4715]: I1203 23:09:55.265599 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-4bvpr_ce5706ce-67d9-4e61-8051-03113e8c3ac4/cert-manager-controller/0.log" Dec 03 23:09:55 crc kubenswrapper[4715]: I1203 23:09:55.421867 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-dgskj_30ac245b-49dd-48b9-878e-e06008f921b7/cert-manager-webhook/0.log" Dec 03 23:09:59 crc kubenswrapper[4715]: I1203 23:09:59.634371 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:09:59 crc kubenswrapper[4715]: E1203 23:09:59.635765 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:10:07 crc kubenswrapper[4715]: I1203 23:10:07.492168 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-lkqpr_47af66c3-ee08-4f79-9e8e-f9e1b3457971/nmstate-console-plugin/0.log" Dec 03 23:10:07 crc kubenswrapper[4715]: I1203 23:10:07.661669 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-vkmgw_be31b811-7008-4d9f-ac8b-4c454e4a8597/nmstate-handler/0.log" Dec 03 23:10:07 crc kubenswrapper[4715]: I1203 23:10:07.708618 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9n62x_6d716f67-d1d1-44c0-b2d4-eeb60e73160f/nmstate-metrics/0.log" Dec 03 23:10:07 crc kubenswrapper[4715]: I1203 23:10:07.741673 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9n62x_6d716f67-d1d1-44c0-b2d4-eeb60e73160f/kube-rbac-proxy/0.log" Dec 03 23:10:07 crc kubenswrapper[4715]: I1203 23:10:07.885137 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-xht5s_7bff12a8-00c5-4a84-85f7-3a0055308b87/nmstate-operator/0.log" Dec 03 23:10:07 crc kubenswrapper[4715]: I1203 23:10:07.908337 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-4d8n9_bef574d3-8639-4a82-aa6e-323a58b9db40/nmstate-webhook/0.log" Dec 03 23:10:10 crc kubenswrapper[4715]: I1203 23:10:10.634252 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:10:11 crc kubenswrapper[4715]: I1203 23:10:11.506161 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"c0bb4558490f79c9a3fb62c9b815e883ad85d7cda56ce37156eb0d1f3137cca5"} Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.319282 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-22qzs_a7a0c0e5-fb39-479e-8731-f81bc225c7fb/kube-rbac-proxy/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.408422 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-22qzs_a7a0c0e5-fb39-479e-8731-f81bc225c7fb/controller/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.595573 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-frr-files/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.732771 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-metrics/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.736809 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-reloader/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.750060 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-frr-files/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.779674 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-reloader/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.937622 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-frr-files/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.972984 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-metrics/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.983011 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-metrics/0.log" Dec 03 23:10:22 crc kubenswrapper[4715]: I1203 23:10:22.992805 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-reloader/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.184069 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-reloader/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.189421 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-frr-files/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.202200 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/controller/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.245136 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/cp-metrics/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.350992 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/frr-metrics/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.394051 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/kube-rbac-proxy/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.417147 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/kube-rbac-proxy-frr/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.570646 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/reloader/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.679419 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-94tb9_ff247a17-7b16-4a5e-adcf-48f9122130eb/frr-k8s-webhook-server/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.837147 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-b57df49df-j4jvw_3483831a-11e2-4881-8447-fc127b2ec983/manager/0.log" Dec 03 23:10:23 crc kubenswrapper[4715]: I1203 23:10:23.954736 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7f9b456564-mq9gn_0b23c7d6-56da-4a46-ab0e-e98665166baa/webhook-server/0.log" Dec 03 23:10:24 crc kubenswrapper[4715]: I1203 23:10:24.111046 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rjg8s_b396984d-e5db-4417-ab0b-c2846cc8c95a/kube-rbac-proxy/0.log" Dec 03 23:10:24 crc kubenswrapper[4715]: I1203 23:10:24.618553 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rjg8s_b396984d-e5db-4417-ab0b-c2846cc8c95a/speaker/0.log" Dec 03 23:10:25 crc kubenswrapper[4715]: I1203 23:10:25.000710 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mvgh5_ae01887b-8d01-4eaf-8214-3436b7f7af87/frr/0.log" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.434203 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-z5bjc"] Dec 03 23:10:37 crc kubenswrapper[4715]: E1203 23:10:37.435139 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f66136-98be-47da-8880-9d8e37f6665a" containerName="container-00" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.435151 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f66136-98be-47da-8880-9d8e37f6665a" containerName="container-00" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.435357 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f66136-98be-47da-8880-9d8e37f6665a" containerName="container-00" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.436995 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.450494 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5bjc"] Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.477315 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-catalog-content\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.477445 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-utilities\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.477532 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr9v2\" (UniqueName: \"kubernetes.io/projected/dabcfd15-614f-4c89-90e7-deaaaf17b778-kube-api-access-hr9v2\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.580493 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-catalog-content\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.580647 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-utilities\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.580729 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr9v2\" (UniqueName: \"kubernetes.io/projected/dabcfd15-614f-4c89-90e7-deaaaf17b778-kube-api-access-hr9v2\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.581128 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-catalog-content\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.581158 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-utilities\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.602199 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr9v2\" (UniqueName: \"kubernetes.io/projected/dabcfd15-614f-4c89-90e7-deaaaf17b778-kube-api-access-hr9v2\") pod \"certified-operators-z5bjc\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:37 crc kubenswrapper[4715]: I1203 23:10:37.757423 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.300570 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-z5bjc"] Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.608018 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/util/0.log" Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.774476 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/util/0.log" Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.801849 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/pull/0.log" Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.804549 4715 generic.go:334] "Generic (PLEG): container finished" podID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerID="209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298" exitCode=0 Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.804602 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5bjc" event={"ID":"dabcfd15-614f-4c89-90e7-deaaaf17b778","Type":"ContainerDied","Data":"209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298"} Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.804645 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5bjc" event={"ID":"dabcfd15-614f-4c89-90e7-deaaaf17b778","Type":"ContainerStarted","Data":"a6c41a5b875a5dd16df558bc2edbbefd1fa5cdece38b0522c04b8d3d17b68092"} Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.806329 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.812379 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/pull/0.log" Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.955201 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/util/0.log" Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.956750 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/pull/0.log" Dec 03 23:10:38 crc kubenswrapper[4715]: I1203 23:10:38.981087 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212f65llr_d93c04f8-0649-47f4-b953-4dba2455c11e/extract/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.149418 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/util/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.291977 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/util/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.322139 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/pull/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.323005 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/pull/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.469079 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/pull/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.473034 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/util/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.518814 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83xb67h_680d52ba-8bcc-496c-9a08-c9b118c8cf76/extract/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.646398 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-utilities/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.818269 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5bjc" event={"ID":"dabcfd15-614f-4c89-90e7-deaaaf17b778","Type":"ContainerStarted","Data":"1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957"} Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.866275 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-content/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.905559 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-utilities/0.log" Dec 03 23:10:39 crc kubenswrapper[4715]: I1203 23:10:39.955432 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-content/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.075691 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-utilities/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.085993 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/extract-content/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.370996 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5bjc_dabcfd15-614f-4c89-90e7-deaaaf17b778/extract-utilities/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.573236 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5bjc_dabcfd15-614f-4c89-90e7-deaaaf17b778/extract-content/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.601918 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5bjc_dabcfd15-614f-4c89-90e7-deaaaf17b778/extract-utilities/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.621449 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5bjc_dabcfd15-614f-4c89-90e7-deaaaf17b778/extract-content/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.687478 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-rvbck_63ed1f29-d7e8-4dc2-b660-dbac35ed94ec/registry-server/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.769156 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5bjc_dabcfd15-614f-4c89-90e7-deaaaf17b778/extract-content/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.818139 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-z5bjc_dabcfd15-614f-4c89-90e7-deaaaf17b778/extract-utilities/0.log" Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.826533 4715 generic.go:334] "Generic (PLEG): container finished" podID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerID="1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957" exitCode=0 Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.826567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5bjc" event={"ID":"dabcfd15-614f-4c89-90e7-deaaaf17b778","Type":"ContainerDied","Data":"1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957"} Dec 03 23:10:40 crc kubenswrapper[4715]: I1203 23:10:40.936554 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-utilities/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.106423 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-content/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.119037 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-content/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.130837 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-utilities/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.277152 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-utilities/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.343059 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/extract-content/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.567715 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-utilities/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.583483 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-ss8p4_62dffbaa-068c-484d-822d-51bb3bfd3105/marketplace-operator/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.855925 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5bjc" event={"ID":"dabcfd15-614f-4c89-90e7-deaaaf17b778","Type":"ContainerStarted","Data":"4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5"} Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.905890 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9nxq5_2d979921-24da-4ea9-bb4a-bb7f921778e0/registry-server/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.940829 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-content/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.965971 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-content/0.log" Dec 03 23:10:41 crc kubenswrapper[4715]: I1203 23:10:41.982183 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-utilities/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.136838 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-utilities/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.146096 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/extract-content/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.250065 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-utilities/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.309603 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-v8gpf_554ea49c-3cc3-42fd-b080-3054daed5166/registry-server/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.439601 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-utilities/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.445001 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-content/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.446807 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-content/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.653617 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-utilities/0.log" Dec 03 23:10:42 crc kubenswrapper[4715]: I1203 23:10:42.688237 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/extract-content/0.log" Dec 03 23:10:43 crc kubenswrapper[4715]: I1203 23:10:43.270349 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4g7ql_c759bed5-8b17-4009-90f1-6422be78653d/registry-server/0.log" Dec 03 23:10:47 crc kubenswrapper[4715]: I1203 23:10:47.759407 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:47 crc kubenswrapper[4715]: I1203 23:10:47.760092 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:47 crc kubenswrapper[4715]: I1203 23:10:47.837022 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:47 crc kubenswrapper[4715]: I1203 23:10:47.873595 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-z5bjc" podStartSLOduration=8.270876795 podStartE2EDuration="10.87357593s" podCreationTimestamp="2025-12-03 23:10:37 +0000 UTC" firstStartedPulling="2025-12-03 23:10:38.806124204 +0000 UTC m=+5335.548834789" lastFinishedPulling="2025-12-03 23:10:41.408823339 +0000 UTC m=+5338.151533924" observedRunningTime="2025-12-03 23:10:41.877941847 +0000 UTC m=+5338.620652442" watchObservedRunningTime="2025-12-03 23:10:47.87357593 +0000 UTC m=+5344.616286525" Dec 03 23:10:47 crc kubenswrapper[4715]: I1203 23:10:47.978557 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:48 crc kubenswrapper[4715]: I1203 23:10:48.081417 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z5bjc"] Dec 03 23:10:49 crc kubenswrapper[4715]: I1203 23:10:49.924646 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-z5bjc" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerName="registry-server" containerID="cri-o://4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5" gracePeriod=2 Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.369008 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.523225 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-catalog-content\") pod \"dabcfd15-614f-4c89-90e7-deaaaf17b778\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.523618 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-utilities\") pod \"dabcfd15-614f-4c89-90e7-deaaaf17b778\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.523780 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr9v2\" (UniqueName: \"kubernetes.io/projected/dabcfd15-614f-4c89-90e7-deaaaf17b778-kube-api-access-hr9v2\") pod \"dabcfd15-614f-4c89-90e7-deaaaf17b778\" (UID: \"dabcfd15-614f-4c89-90e7-deaaaf17b778\") " Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.525816 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-utilities" (OuterVolumeSpecName: "utilities") pod "dabcfd15-614f-4c89-90e7-deaaaf17b778" (UID: "dabcfd15-614f-4c89-90e7-deaaaf17b778"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.529126 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dabcfd15-614f-4c89-90e7-deaaaf17b778-kube-api-access-hr9v2" (OuterVolumeSpecName: "kube-api-access-hr9v2") pod "dabcfd15-614f-4c89-90e7-deaaaf17b778" (UID: "dabcfd15-614f-4c89-90e7-deaaaf17b778"). InnerVolumeSpecName "kube-api-access-hr9v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.626249 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.626293 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr9v2\" (UniqueName: \"kubernetes.io/projected/dabcfd15-614f-4c89-90e7-deaaaf17b778-kube-api-access-hr9v2\") on node \"crc\" DevicePath \"\"" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.678115 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dabcfd15-614f-4c89-90e7-deaaaf17b778" (UID: "dabcfd15-614f-4c89-90e7-deaaaf17b778"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.728762 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dabcfd15-614f-4c89-90e7-deaaaf17b778-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.936072 4715 generic.go:334] "Generic (PLEG): container finished" podID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerID="4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5" exitCode=0 Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.936124 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5bjc" event={"ID":"dabcfd15-614f-4c89-90e7-deaaaf17b778","Type":"ContainerDied","Data":"4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5"} Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.936157 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-z5bjc" event={"ID":"dabcfd15-614f-4c89-90e7-deaaaf17b778","Type":"ContainerDied","Data":"a6c41a5b875a5dd16df558bc2edbbefd1fa5cdece38b0522c04b8d3d17b68092"} Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.936180 4715 scope.go:117] "RemoveContainer" containerID="4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.936318 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-z5bjc" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.982073 4715 scope.go:117] "RemoveContainer" containerID="1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957" Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.986285 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-z5bjc"] Dec 03 23:10:50 crc kubenswrapper[4715]: I1203 23:10:50.999379 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-z5bjc"] Dec 03 23:10:51 crc kubenswrapper[4715]: I1203 23:10:51.007751 4715 scope.go:117] "RemoveContainer" containerID="209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298" Dec 03 23:10:51 crc kubenswrapper[4715]: I1203 23:10:51.075259 4715 scope.go:117] "RemoveContainer" containerID="4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5" Dec 03 23:10:51 crc kubenswrapper[4715]: E1203 23:10:51.075966 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5\": container with ID starting with 4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5 not found: ID does not exist" containerID="4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5" Dec 03 23:10:51 crc kubenswrapper[4715]: I1203 23:10:51.076005 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5"} err="failed to get container status \"4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5\": rpc error: code = NotFound desc = could not find container \"4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5\": container with ID starting with 4a337eeb973f47d5b642a721f22a16408149b3aab9b25d6370db052ba7813fc5 not found: ID does not exist" Dec 03 23:10:51 crc kubenswrapper[4715]: I1203 23:10:51.076029 4715 scope.go:117] "RemoveContainer" containerID="1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957" Dec 03 23:10:51 crc kubenswrapper[4715]: E1203 23:10:51.076588 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957\": container with ID starting with 1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957 not found: ID does not exist" containerID="1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957" Dec 03 23:10:51 crc kubenswrapper[4715]: I1203 23:10:51.076622 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957"} err="failed to get container status \"1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957\": rpc error: code = NotFound desc = could not find container \"1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957\": container with ID starting with 1c85870de05819002ffb6d002710c047d12fa0646b3fb698f413ed5c76600957 not found: ID does not exist" Dec 03 23:10:51 crc kubenswrapper[4715]: I1203 23:10:51.076643 4715 scope.go:117] "RemoveContainer" containerID="209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298" Dec 03 23:10:51 crc kubenswrapper[4715]: E1203 23:10:51.077164 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298\": container with ID starting with 209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298 not found: ID does not exist" containerID="209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298" Dec 03 23:10:51 crc kubenswrapper[4715]: I1203 23:10:51.077222 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298"} err="failed to get container status \"209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298\": rpc error: code = NotFound desc = could not find container \"209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298\": container with ID starting with 209fcdfabe4f2bdf0a6fa1db4486d7c1db5e0462adcc0bbb5072010d16ff9298 not found: ID does not exist" Dec 03 23:10:51 crc kubenswrapper[4715]: I1203 23:10:51.643799 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" path="/var/lib/kubelet/pods/dabcfd15-614f-4c89-90e7-deaaaf17b778/volumes" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.372234 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b9glq"] Dec 03 23:11:38 crc kubenswrapper[4715]: E1203 23:11:38.373287 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerName="registry-server" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.373305 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerName="registry-server" Dec 03 23:11:38 crc kubenswrapper[4715]: E1203 23:11:38.373327 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerName="extract-utilities" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.373337 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerName="extract-utilities" Dec 03 23:11:38 crc kubenswrapper[4715]: E1203 23:11:38.373354 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerName="extract-content" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.373362 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerName="extract-content" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.373617 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="dabcfd15-614f-4c89-90e7-deaaaf17b778" containerName="registry-server" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.375456 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.387698 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9glq"] Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.395916 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-utilities\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.395997 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-catalog-content\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.396036 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqlt8\" (UniqueName: \"kubernetes.io/projected/d871fdb4-0387-459d-bcd6-441b9d96bc3f-kube-api-access-jqlt8\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.497283 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-utilities\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.497347 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-catalog-content\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.497374 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqlt8\" (UniqueName: \"kubernetes.io/projected/d871fdb4-0387-459d-bcd6-441b9d96bc3f-kube-api-access-jqlt8\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.498209 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-utilities\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.498432 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-catalog-content\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.517205 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqlt8\" (UniqueName: \"kubernetes.io/projected/d871fdb4-0387-459d-bcd6-441b9d96bc3f-kube-api-access-jqlt8\") pod \"redhat-marketplace-b9glq\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:38 crc kubenswrapper[4715]: I1203 23:11:38.720994 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:39 crc kubenswrapper[4715]: I1203 23:11:39.294698 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9glq"] Dec 03 23:11:39 crc kubenswrapper[4715]: W1203 23:11:39.301006 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd871fdb4_0387_459d_bcd6_441b9d96bc3f.slice/crio-466396eae5bfb0e1a2f5187a4f6cb346b86f9d0f31206b9a82643c0a6d54cb80 WatchSource:0}: Error finding container 466396eae5bfb0e1a2f5187a4f6cb346b86f9d0f31206b9a82643c0a6d54cb80: Status 404 returned error can't find the container with id 466396eae5bfb0e1a2f5187a4f6cb346b86f9d0f31206b9a82643c0a6d54cb80 Dec 03 23:11:39 crc kubenswrapper[4715]: I1203 23:11:39.468018 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9glq" event={"ID":"d871fdb4-0387-459d-bcd6-441b9d96bc3f","Type":"ContainerStarted","Data":"466396eae5bfb0e1a2f5187a4f6cb346b86f9d0f31206b9a82643c0a6d54cb80"} Dec 03 23:11:40 crc kubenswrapper[4715]: I1203 23:11:40.482380 4715 generic.go:334] "Generic (PLEG): container finished" podID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerID="eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf" exitCode=0 Dec 03 23:11:40 crc kubenswrapper[4715]: I1203 23:11:40.482526 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9glq" event={"ID":"d871fdb4-0387-459d-bcd6-441b9d96bc3f","Type":"ContainerDied","Data":"eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf"} Dec 03 23:11:41 crc kubenswrapper[4715]: I1203 23:11:41.496567 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9glq" event={"ID":"d871fdb4-0387-459d-bcd6-441b9d96bc3f","Type":"ContainerStarted","Data":"213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee"} Dec 03 23:11:42 crc kubenswrapper[4715]: I1203 23:11:42.514735 4715 generic.go:334] "Generic (PLEG): container finished" podID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerID="213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee" exitCode=0 Dec 03 23:11:42 crc kubenswrapper[4715]: I1203 23:11:42.514838 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9glq" event={"ID":"d871fdb4-0387-459d-bcd6-441b9d96bc3f","Type":"ContainerDied","Data":"213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee"} Dec 03 23:11:43 crc kubenswrapper[4715]: I1203 23:11:43.529098 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9glq" event={"ID":"d871fdb4-0387-459d-bcd6-441b9d96bc3f","Type":"ContainerStarted","Data":"f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3"} Dec 03 23:11:43 crc kubenswrapper[4715]: I1203 23:11:43.555617 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b9glq" podStartSLOduration=3.11684787 podStartE2EDuration="5.55559102s" podCreationTimestamp="2025-12-03 23:11:38 +0000 UTC" firstStartedPulling="2025-12-03 23:11:40.486465279 +0000 UTC m=+5397.229175914" lastFinishedPulling="2025-12-03 23:11:42.925208459 +0000 UTC m=+5399.667919064" observedRunningTime="2025-12-03 23:11:43.554581043 +0000 UTC m=+5400.297291658" watchObservedRunningTime="2025-12-03 23:11:43.55559102 +0000 UTC m=+5400.298301625" Dec 03 23:11:48 crc kubenswrapper[4715]: I1203 23:11:48.722306 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:48 crc kubenswrapper[4715]: I1203 23:11:48.723695 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:48 crc kubenswrapper[4715]: I1203 23:11:48.812652 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:49 crc kubenswrapper[4715]: I1203 23:11:49.664514 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:49 crc kubenswrapper[4715]: I1203 23:11:49.724363 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9glq"] Dec 03 23:11:51 crc kubenswrapper[4715]: I1203 23:11:51.621436 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b9glq" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerName="registry-server" containerID="cri-o://f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3" gracePeriod=2 Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.261715 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.400314 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqlt8\" (UniqueName: \"kubernetes.io/projected/d871fdb4-0387-459d-bcd6-441b9d96bc3f-kube-api-access-jqlt8\") pod \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.400444 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-catalog-content\") pod \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.400585 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-utilities\") pod \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\" (UID: \"d871fdb4-0387-459d-bcd6-441b9d96bc3f\") " Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.401720 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-utilities" (OuterVolumeSpecName: "utilities") pod "d871fdb4-0387-459d-bcd6-441b9d96bc3f" (UID: "d871fdb4-0387-459d-bcd6-441b9d96bc3f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.407909 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d871fdb4-0387-459d-bcd6-441b9d96bc3f-kube-api-access-jqlt8" (OuterVolumeSpecName: "kube-api-access-jqlt8") pod "d871fdb4-0387-459d-bcd6-441b9d96bc3f" (UID: "d871fdb4-0387-459d-bcd6-441b9d96bc3f"). InnerVolumeSpecName "kube-api-access-jqlt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.435208 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d871fdb4-0387-459d-bcd6-441b9d96bc3f" (UID: "d871fdb4-0387-459d-bcd6-441b9d96bc3f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.503106 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqlt8\" (UniqueName: \"kubernetes.io/projected/d871fdb4-0387-459d-bcd6-441b9d96bc3f-kube-api-access-jqlt8\") on node \"crc\" DevicePath \"\"" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.503146 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.503161 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d871fdb4-0387-459d-bcd6-441b9d96bc3f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.637232 4715 generic.go:334] "Generic (PLEG): container finished" podID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerID="f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3" exitCode=0 Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.637312 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9glq" event={"ID":"d871fdb4-0387-459d-bcd6-441b9d96bc3f","Type":"ContainerDied","Data":"f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3"} Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.637366 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b9glq" event={"ID":"d871fdb4-0387-459d-bcd6-441b9d96bc3f","Type":"ContainerDied","Data":"466396eae5bfb0e1a2f5187a4f6cb346b86f9d0f31206b9a82643c0a6d54cb80"} Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.637410 4715 scope.go:117] "RemoveContainer" containerID="f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.637694 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b9glq" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.664261 4715 scope.go:117] "RemoveContainer" containerID="213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.696369 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9glq"] Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.704806 4715 scope.go:117] "RemoveContainer" containerID="eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.715398 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b9glq"] Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.754704 4715 scope.go:117] "RemoveContainer" containerID="f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3" Dec 03 23:11:52 crc kubenswrapper[4715]: E1203 23:11:52.755327 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3\": container with ID starting with f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3 not found: ID does not exist" containerID="f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.755376 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3"} err="failed to get container status \"f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3\": rpc error: code = NotFound desc = could not find container \"f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3\": container with ID starting with f6efc74919406c0636db9c2c97a947018d19379be5171e086329eacca1afcea3 not found: ID does not exist" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.755409 4715 scope.go:117] "RemoveContainer" containerID="213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee" Dec 03 23:11:52 crc kubenswrapper[4715]: E1203 23:11:52.756236 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee\": container with ID starting with 213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee not found: ID does not exist" containerID="213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.756279 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee"} err="failed to get container status \"213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee\": rpc error: code = NotFound desc = could not find container \"213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee\": container with ID starting with 213d9531bf530247e54bf79d1801db035f9244fb74b35ebf55bbb153df83a3ee not found: ID does not exist" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.756312 4715 scope.go:117] "RemoveContainer" containerID="eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf" Dec 03 23:11:52 crc kubenswrapper[4715]: E1203 23:11:52.756703 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf\": container with ID starting with eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf not found: ID does not exist" containerID="eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf" Dec 03 23:11:52 crc kubenswrapper[4715]: I1203 23:11:52.756759 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf"} err="failed to get container status \"eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf\": rpc error: code = NotFound desc = could not find container \"eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf\": container with ID starting with eaf188accf5e7ff2901f385af8205d97cab27d31dd84efdf54436c1dd78790bf not found: ID does not exist" Dec 03 23:11:53 crc kubenswrapper[4715]: I1203 23:11:53.651106 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" path="/var/lib/kubelet/pods/d871fdb4-0387-459d-bcd6-441b9d96bc3f/volumes" Dec 03 23:12:18 crc kubenswrapper[4715]: I1203 23:12:18.927885 4715 generic.go:334] "Generic (PLEG): container finished" podID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerID="33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1" exitCode=0 Dec 03 23:12:18 crc kubenswrapper[4715]: I1203 23:12:18.928995 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-btdhv/must-gather-pc8sw" event={"ID":"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b","Type":"ContainerDied","Data":"33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1"} Dec 03 23:12:18 crc kubenswrapper[4715]: I1203 23:12:18.930108 4715 scope.go:117] "RemoveContainer" containerID="33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1" Dec 03 23:12:19 crc kubenswrapper[4715]: I1203 23:12:19.572433 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-btdhv_must-gather-pc8sw_83d2f2fb-05a7-4b25-b13f-7bc101c25b8b/gather/0.log" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.064213 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nrp52"] Dec 03 23:12:29 crc kubenswrapper[4715]: E1203 23:12:29.065387 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerName="extract-content" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.065408 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerName="extract-content" Dec 03 23:12:29 crc kubenswrapper[4715]: E1203 23:12:29.065453 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerName="registry-server" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.065466 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerName="registry-server" Dec 03 23:12:29 crc kubenswrapper[4715]: E1203 23:12:29.065526 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerName="extract-utilities" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.065540 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerName="extract-utilities" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.065876 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="d871fdb4-0387-459d-bcd6-441b9d96bc3f" containerName="registry-server" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.068241 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.076075 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nrp52"] Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.078263 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqqzq\" (UniqueName: \"kubernetes.io/projected/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-kube-api-access-fqqzq\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.078310 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-utilities\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.078372 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-catalog-content\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.181299 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqqzq\" (UniqueName: \"kubernetes.io/projected/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-kube-api-access-fqqzq\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.181368 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-utilities\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.181435 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-catalog-content\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.181992 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-catalog-content\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.182038 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-utilities\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.202634 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqqzq\" (UniqueName: \"kubernetes.io/projected/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-kube-api-access-fqqzq\") pod \"community-operators-nrp52\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.399135 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:29 crc kubenswrapper[4715]: I1203 23:12:29.887978 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nrp52"] Dec 03 23:12:29 crc kubenswrapper[4715]: W1203 23:12:29.894915 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda225f3a9_d3ed_481e_a6ea_f97d78a2d7fb.slice/crio-051128b9c469fb7a636239afc0ad653f69ab9502de256581e1466af6f37e2b95 WatchSource:0}: Error finding container 051128b9c469fb7a636239afc0ad653f69ab9502de256581e1466af6f37e2b95: Status 404 returned error can't find the container with id 051128b9c469fb7a636239afc0ad653f69ab9502de256581e1466af6f37e2b95 Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.033426 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-btdhv/must-gather-pc8sw"] Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.033758 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-btdhv/must-gather-pc8sw" podUID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerName="copy" containerID="cri-o://393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127" gracePeriod=2 Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.044494 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-btdhv/must-gather-pc8sw"] Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.064994 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrp52" event={"ID":"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb","Type":"ContainerStarted","Data":"051128b9c469fb7a636239afc0ad653f69ab9502de256581e1466af6f37e2b95"} Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.509741 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-btdhv_must-gather-pc8sw_83d2f2fb-05a7-4b25-b13f-7bc101c25b8b/copy/0.log" Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.510433 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.711079 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-must-gather-output\") pod \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\" (UID: \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\") " Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.711166 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k9w2\" (UniqueName: \"kubernetes.io/projected/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-kube-api-access-2k9w2\") pod \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\" (UID: \"83d2f2fb-05a7-4b25-b13f-7bc101c25b8b\") " Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.717791 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-kube-api-access-2k9w2" (OuterVolumeSpecName: "kube-api-access-2k9w2") pod "83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" (UID: "83d2f2fb-05a7-4b25-b13f-7bc101c25b8b"). InnerVolumeSpecName "kube-api-access-2k9w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.816858 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k9w2\" (UniqueName: \"kubernetes.io/projected/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-kube-api-access-2k9w2\") on node \"crc\" DevicePath \"\"" Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.891707 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" (UID: "83d2f2fb-05a7-4b25-b13f-7bc101c25b8b"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:12:30 crc kubenswrapper[4715]: I1203 23:12:30.920306 4715 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.074620 4715 generic.go:334] "Generic (PLEG): container finished" podID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerID="225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9" exitCode=0 Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.074675 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrp52" event={"ID":"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb","Type":"ContainerDied","Data":"225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9"} Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.076616 4715 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-btdhv_must-gather-pc8sw_83d2f2fb-05a7-4b25-b13f-7bc101c25b8b/copy/0.log" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.076997 4715 generic.go:334] "Generic (PLEG): container finished" podID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerID="393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127" exitCode=143 Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.077042 4715 scope.go:117] "RemoveContainer" containerID="393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.077070 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-btdhv/must-gather-pc8sw" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.096705 4715 scope.go:117] "RemoveContainer" containerID="33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.191097 4715 scope.go:117] "RemoveContainer" containerID="393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127" Dec 03 23:12:31 crc kubenswrapper[4715]: E1203 23:12:31.191544 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127\": container with ID starting with 393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127 not found: ID does not exist" containerID="393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.191578 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127"} err="failed to get container status \"393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127\": rpc error: code = NotFound desc = could not find container \"393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127\": container with ID starting with 393e1ac8e083406ca39d5c27644a773487d47f5a29494dd2231cb98248617127 not found: ID does not exist" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.191597 4715 scope.go:117] "RemoveContainer" containerID="33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1" Dec 03 23:12:31 crc kubenswrapper[4715]: E1203 23:12:31.191842 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1\": container with ID starting with 33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1 not found: ID does not exist" containerID="33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.191869 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1"} err="failed to get container status \"33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1\": rpc error: code = NotFound desc = could not find container \"33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1\": container with ID starting with 33ec1358b8094a29518f90028ac3d8fcf9e39b390044493e7d79a958a51630f1 not found: ID does not exist" Dec 03 23:12:31 crc kubenswrapper[4715]: I1203 23:12:31.653083 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" path="/var/lib/kubelet/pods/83d2f2fb-05a7-4b25-b13f-7bc101c25b8b/volumes" Dec 03 23:12:32 crc kubenswrapper[4715]: I1203 23:12:32.090526 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrp52" event={"ID":"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb","Type":"ContainerStarted","Data":"6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377"} Dec 03 23:12:33 crc kubenswrapper[4715]: I1203 23:12:33.104325 4715 generic.go:334] "Generic (PLEG): container finished" podID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerID="6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377" exitCode=0 Dec 03 23:12:33 crc kubenswrapper[4715]: I1203 23:12:33.104420 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrp52" event={"ID":"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb","Type":"ContainerDied","Data":"6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377"} Dec 03 23:12:34 crc kubenswrapper[4715]: I1203 23:12:34.116237 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrp52" event={"ID":"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb","Type":"ContainerStarted","Data":"e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0"} Dec 03 23:12:34 crc kubenswrapper[4715]: I1203 23:12:34.156735 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nrp52" podStartSLOduration=2.696063721 podStartE2EDuration="5.156708932s" podCreationTimestamp="2025-12-03 23:12:29 +0000 UTC" firstStartedPulling="2025-12-03 23:12:31.076242705 +0000 UTC m=+5447.818953300" lastFinishedPulling="2025-12-03 23:12:33.536887896 +0000 UTC m=+5450.279598511" observedRunningTime="2025-12-03 23:12:34.142579391 +0000 UTC m=+5450.885290006" watchObservedRunningTime="2025-12-03 23:12:34.156708932 +0000 UTC m=+5450.899419567" Dec 03 23:12:35 crc kubenswrapper[4715]: I1203 23:12:35.159768 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:12:35 crc kubenswrapper[4715]: I1203 23:12:35.160094 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:12:39 crc kubenswrapper[4715]: I1203 23:12:39.399934 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:39 crc kubenswrapper[4715]: I1203 23:12:39.400560 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:39 crc kubenswrapper[4715]: I1203 23:12:39.483037 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:40 crc kubenswrapper[4715]: I1203 23:12:40.264352 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:40 crc kubenswrapper[4715]: I1203 23:12:40.325049 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nrp52"] Dec 03 23:12:42 crc kubenswrapper[4715]: I1203 23:12:42.202784 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nrp52" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerName="registry-server" containerID="cri-o://e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0" gracePeriod=2 Dec 03 23:12:42 crc kubenswrapper[4715]: I1203 23:12:42.766189 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:42 crc kubenswrapper[4715]: I1203 23:12:42.957237 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-utilities\") pod \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " Dec 03 23:12:42 crc kubenswrapper[4715]: I1203 23:12:42.957528 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqqzq\" (UniqueName: \"kubernetes.io/projected/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-kube-api-access-fqqzq\") pod \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " Dec 03 23:12:42 crc kubenswrapper[4715]: I1203 23:12:42.957613 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-catalog-content\") pod \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\" (UID: \"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb\") " Dec 03 23:12:42 crc kubenswrapper[4715]: I1203 23:12:42.958791 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-utilities" (OuterVolumeSpecName: "utilities") pod "a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" (UID: "a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:12:42 crc kubenswrapper[4715]: I1203 23:12:42.965780 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-kube-api-access-fqqzq" (OuterVolumeSpecName: "kube-api-access-fqqzq") pod "a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" (UID: "a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb"). InnerVolumeSpecName "kube-api-access-fqqzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.060467 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.060530 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqqzq\" (UniqueName: \"kubernetes.io/projected/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-kube-api-access-fqqzq\") on node \"crc\" DevicePath \"\"" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.157300 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" (UID: "a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.162126 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.218868 4715 generic.go:334] "Generic (PLEG): container finished" podID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerID="e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0" exitCode=0 Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.218924 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrp52" event={"ID":"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb","Type":"ContainerDied","Data":"e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0"} Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.218955 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nrp52" event={"ID":"a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb","Type":"ContainerDied","Data":"051128b9c469fb7a636239afc0ad653f69ab9502de256581e1466af6f37e2b95"} Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.218989 4715 scope.go:117] "RemoveContainer" containerID="e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.219317 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nrp52" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.257902 4715 scope.go:117] "RemoveContainer" containerID="6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.262042 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nrp52"] Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.273184 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nrp52"] Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.282694 4715 scope.go:117] "RemoveContainer" containerID="225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.327936 4715 scope.go:117] "RemoveContainer" containerID="e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0" Dec 03 23:12:43 crc kubenswrapper[4715]: E1203 23:12:43.328484 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0\": container with ID starting with e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0 not found: ID does not exist" containerID="e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.328580 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0"} err="failed to get container status \"e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0\": rpc error: code = NotFound desc = could not find container \"e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0\": container with ID starting with e3c2edacf5dbe902bc21f1cf14f898a63926d1d3369e03c5ec008e46d24792f0 not found: ID does not exist" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.328620 4715 scope.go:117] "RemoveContainer" containerID="6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377" Dec 03 23:12:43 crc kubenswrapper[4715]: E1203 23:12:43.329376 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377\": container with ID starting with 6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377 not found: ID does not exist" containerID="6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.329441 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377"} err="failed to get container status \"6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377\": rpc error: code = NotFound desc = could not find container \"6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377\": container with ID starting with 6b84264c8764f66000751b0267d753348de9185d93d0677e98c31f82d0987377 not found: ID does not exist" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.329475 4715 scope.go:117] "RemoveContainer" containerID="225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9" Dec 03 23:12:43 crc kubenswrapper[4715]: E1203 23:12:43.329875 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9\": container with ID starting with 225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9 not found: ID does not exist" containerID="225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.329911 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9"} err="failed to get container status \"225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9\": rpc error: code = NotFound desc = could not find container \"225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9\": container with ID starting with 225ac976369fc7909733541990624795dda1b88ef074ae96ae1d92675c0835e9 not found: ID does not exist" Dec 03 23:12:43 crc kubenswrapper[4715]: I1203 23:12:43.646393 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" path="/var/lib/kubelet/pods/a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb/volumes" Dec 03 23:13:05 crc kubenswrapper[4715]: I1203 23:13:05.160200 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:13:05 crc kubenswrapper[4715]: I1203 23:13:05.160702 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.160482 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.161175 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.161251 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.162180 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c0bb4558490f79c9a3fb62c9b815e883ad85d7cda56ce37156eb0d1f3137cca5"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.162277 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://c0bb4558490f79c9a3fb62c9b815e883ad85d7cda56ce37156eb0d1f3137cca5" gracePeriod=600 Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.821962 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="c0bb4558490f79c9a3fb62c9b815e883ad85d7cda56ce37156eb0d1f3137cca5" exitCode=0 Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.822046 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"c0bb4558490f79c9a3fb62c9b815e883ad85d7cda56ce37156eb0d1f3137cca5"} Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.822380 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerStarted","Data":"ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15"} Dec 03 23:13:35 crc kubenswrapper[4715]: I1203 23:13:35.822408 4715 scope.go:117] "RemoveContainer" containerID="6dee6856420c4db6f492cf11e5c64e14533d171c83daaa0dea1aa5261d1d481b" Dec 03 23:13:57 crc kubenswrapper[4715]: I1203 23:13:57.263333 4715 scope.go:117] "RemoveContainer" containerID="95d857137edda94c308409535df37e7704d5cd1435136a1e7f9ce3df6511422d" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.155476 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7"] Dec 03 23:15:00 crc kubenswrapper[4715]: E1203 23:15:00.156423 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerName="registry-server" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.156440 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerName="registry-server" Dec 03 23:15:00 crc kubenswrapper[4715]: E1203 23:15:00.156465 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerName="extract-utilities" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.156472 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerName="extract-utilities" Dec 03 23:15:00 crc kubenswrapper[4715]: E1203 23:15:00.156489 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerName="extract-content" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.156517 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerName="extract-content" Dec 03 23:15:00 crc kubenswrapper[4715]: E1203 23:15:00.156551 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerName="copy" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.156560 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerName="copy" Dec 03 23:15:00 crc kubenswrapper[4715]: E1203 23:15:00.156568 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerName="gather" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.156574 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerName="gather" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.156797 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="a225f3a9-d3ed-481e-a6ea-f97d78a2d7fb" containerName="registry-server" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.156817 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerName="copy" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.156832 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d2f2fb-05a7-4b25-b13f-7bc101c25b8b" containerName="gather" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.157722 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.160614 4715 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.160628 4715 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.182312 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7"] Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.212390 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f8e362a-cac6-4168-b41c-a5612f07f49f-secret-volume\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.212459 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-456dg\" (UniqueName: \"kubernetes.io/projected/5f8e362a-cac6-4168-b41c-a5612f07f49f-kube-api-access-456dg\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.212556 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f8e362a-cac6-4168-b41c-a5612f07f49f-config-volume\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.314933 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f8e362a-cac6-4168-b41c-a5612f07f49f-secret-volume\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.315258 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-456dg\" (UniqueName: \"kubernetes.io/projected/5f8e362a-cac6-4168-b41c-a5612f07f49f-kube-api-access-456dg\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.315398 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f8e362a-cac6-4168-b41c-a5612f07f49f-config-volume\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.316438 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f8e362a-cac6-4168-b41c-a5612f07f49f-config-volume\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.323572 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f8e362a-cac6-4168-b41c-a5612f07f49f-secret-volume\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.337273 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-456dg\" (UniqueName: \"kubernetes.io/projected/5f8e362a-cac6-4168-b41c-a5612f07f49f-kube-api-access-456dg\") pod \"collect-profiles-29413395-cbzp7\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.487993 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:00 crc kubenswrapper[4715]: I1203 23:15:00.950978 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7"] Dec 03 23:15:00 crc kubenswrapper[4715]: W1203 23:15:00.959433 4715 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f8e362a_cac6_4168_b41c_a5612f07f49f.slice/crio-4ef58e140f301ca2298500ef04259aeda9a05d5821c63f0515ce2aa8921242ca WatchSource:0}: Error finding container 4ef58e140f301ca2298500ef04259aeda9a05d5821c63f0515ce2aa8921242ca: Status 404 returned error can't find the container with id 4ef58e140f301ca2298500ef04259aeda9a05d5821c63f0515ce2aa8921242ca Dec 03 23:15:01 crc kubenswrapper[4715]: I1203 23:15:01.783716 4715 generic.go:334] "Generic (PLEG): container finished" podID="5f8e362a-cac6-4168-b41c-a5612f07f49f" containerID="e32d6b6c2fad4b1ee4008e33f0be97d764c796610431f0e789fed1f0f0d5f4d0" exitCode=0 Dec 03 23:15:01 crc kubenswrapper[4715]: I1203 23:15:01.783779 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" event={"ID":"5f8e362a-cac6-4168-b41c-a5612f07f49f","Type":"ContainerDied","Data":"e32d6b6c2fad4b1ee4008e33f0be97d764c796610431f0e789fed1f0f0d5f4d0"} Dec 03 23:15:01 crc kubenswrapper[4715]: I1203 23:15:01.783955 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" event={"ID":"5f8e362a-cac6-4168-b41c-a5612f07f49f","Type":"ContainerStarted","Data":"4ef58e140f301ca2298500ef04259aeda9a05d5821c63f0515ce2aa8921242ca"} Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.143017 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.291674 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f8e362a-cac6-4168-b41c-a5612f07f49f-config-volume\") pod \"5f8e362a-cac6-4168-b41c-a5612f07f49f\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.291809 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f8e362a-cac6-4168-b41c-a5612f07f49f-secret-volume\") pod \"5f8e362a-cac6-4168-b41c-a5612f07f49f\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.291974 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-456dg\" (UniqueName: \"kubernetes.io/projected/5f8e362a-cac6-4168-b41c-a5612f07f49f-kube-api-access-456dg\") pod \"5f8e362a-cac6-4168-b41c-a5612f07f49f\" (UID: \"5f8e362a-cac6-4168-b41c-a5612f07f49f\") " Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.292720 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f8e362a-cac6-4168-b41c-a5612f07f49f-config-volume" (OuterVolumeSpecName: "config-volume") pod "5f8e362a-cac6-4168-b41c-a5612f07f49f" (UID: "5f8e362a-cac6-4168-b41c-a5612f07f49f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.298117 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f8e362a-cac6-4168-b41c-a5612f07f49f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5f8e362a-cac6-4168-b41c-a5612f07f49f" (UID: "5f8e362a-cac6-4168-b41c-a5612f07f49f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.298942 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f8e362a-cac6-4168-b41c-a5612f07f49f-kube-api-access-456dg" (OuterVolumeSpecName: "kube-api-access-456dg") pod "5f8e362a-cac6-4168-b41c-a5612f07f49f" (UID: "5f8e362a-cac6-4168-b41c-a5612f07f49f"). InnerVolumeSpecName "kube-api-access-456dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.394154 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-456dg\" (UniqueName: \"kubernetes.io/projected/5f8e362a-cac6-4168-b41c-a5612f07f49f-kube-api-access-456dg\") on node \"crc\" DevicePath \"\"" Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.394204 4715 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5f8e362a-cac6-4168-b41c-a5612f07f49f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.394222 4715 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5f8e362a-cac6-4168-b41c-a5612f07f49f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.810389 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" event={"ID":"5f8e362a-cac6-4168-b41c-a5612f07f49f","Type":"ContainerDied","Data":"4ef58e140f301ca2298500ef04259aeda9a05d5821c63f0515ce2aa8921242ca"} Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.810444 4715 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ef58e140f301ca2298500ef04259aeda9a05d5821c63f0515ce2aa8921242ca" Dec 03 23:15:03 crc kubenswrapper[4715]: I1203 23:15:03.810607 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413395-cbzp7" Dec 03 23:15:04 crc kubenswrapper[4715]: I1203 23:15:04.224519 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr"] Dec 03 23:15:04 crc kubenswrapper[4715]: I1203 23:15:04.233155 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413350-4h5tr"] Dec 03 23:15:05 crc kubenswrapper[4715]: I1203 23:15:05.650648 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="633f03e8-4172-4791-8e99-c7f31ae34dba" path="/var/lib/kubelet/pods/633f03e8-4172-4791-8e99-c7f31ae34dba/volumes" Dec 03 23:15:35 crc kubenswrapper[4715]: I1203 23:15:35.159791 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:15:35 crc kubenswrapper[4715]: I1203 23:15:35.160416 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:15:57 crc kubenswrapper[4715]: I1203 23:15:57.344904 4715 scope.go:117] "RemoveContainer" containerID="b9eeeb7de6354154d758d2eb5bd6db2cb6425df7fd02c47d4efa8e091f685a2c" Dec 03 23:16:05 crc kubenswrapper[4715]: I1203 23:16:05.159822 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:16:05 crc kubenswrapper[4715]: I1203 23:16:05.160375 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.160060 4715 patch_prober.go:28] interesting pod/machine-config-daemon-ctls2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.160682 4715 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.160744 4715 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.161754 4715 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15"} pod="openshift-machine-config-operator/machine-config-daemon-ctls2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.161848 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerName="machine-config-daemon" containerID="cri-o://ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" gracePeriod=600 Dec 03 23:16:35 crc kubenswrapper[4715]: E1203 23:16:35.282498 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.829249 4715 generic.go:334] "Generic (PLEG): container finished" podID="0ca0ce89-c3ec-421f-9954-c594760b16a9" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" exitCode=0 Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.829295 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" event={"ID":"0ca0ce89-c3ec-421f-9954-c594760b16a9","Type":"ContainerDied","Data":"ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15"} Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.829819 4715 scope.go:117] "RemoveContainer" containerID="c0bb4558490f79c9a3fb62c9b815e883ad85d7cda56ce37156eb0d1f3137cca5" Dec 03 23:16:35 crc kubenswrapper[4715]: I1203 23:16:35.830888 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:16:35 crc kubenswrapper[4715]: E1203 23:16:35.831453 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:16:46 crc kubenswrapper[4715]: I1203 23:16:46.634888 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:16:46 crc kubenswrapper[4715]: E1203 23:16:46.636176 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:17:01 crc kubenswrapper[4715]: I1203 23:17:01.634667 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:17:01 crc kubenswrapper[4715]: E1203 23:17:01.635477 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:17:14 crc kubenswrapper[4715]: I1203 23:17:14.633986 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:17:14 crc kubenswrapper[4715]: E1203 23:17:14.634767 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:17:27 crc kubenswrapper[4715]: I1203 23:17:27.786737 4715 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wtjn8"] Dec 03 23:17:27 crc kubenswrapper[4715]: E1203 23:17:27.787953 4715 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f8e362a-cac6-4168-b41c-a5612f07f49f" containerName="collect-profiles" Dec 03 23:17:27 crc kubenswrapper[4715]: I1203 23:17:27.787974 4715 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f8e362a-cac6-4168-b41c-a5612f07f49f" containerName="collect-profiles" Dec 03 23:17:27 crc kubenswrapper[4715]: I1203 23:17:27.788288 4715 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f8e362a-cac6-4168-b41c-a5612f07f49f" containerName="collect-profiles" Dec 03 23:17:27 crc kubenswrapper[4715]: I1203 23:17:27.790655 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:27 crc kubenswrapper[4715]: I1203 23:17:27.797563 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wtjn8"] Dec 03 23:17:27 crc kubenswrapper[4715]: I1203 23:17:27.922797 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kpvm\" (UniqueName: \"kubernetes.io/projected/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-kube-api-access-2kpvm\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:27 crc kubenswrapper[4715]: I1203 23:17:27.923065 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-catalog-content\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:27 crc kubenswrapper[4715]: I1203 23:17:27.923158 4715 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-utilities\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:28 crc kubenswrapper[4715]: I1203 23:17:28.025162 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-utilities\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:28 crc kubenswrapper[4715]: I1203 23:17:28.025288 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kpvm\" (UniqueName: \"kubernetes.io/projected/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-kube-api-access-2kpvm\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:28 crc kubenswrapper[4715]: I1203 23:17:28.025350 4715 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-catalog-content\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:28 crc kubenswrapper[4715]: I1203 23:17:28.025804 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-utilities\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:28 crc kubenswrapper[4715]: I1203 23:17:28.025825 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-catalog-content\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:28 crc kubenswrapper[4715]: I1203 23:17:28.048279 4715 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kpvm\" (UniqueName: \"kubernetes.io/projected/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-kube-api-access-2kpvm\") pod \"redhat-operators-wtjn8\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:28 crc kubenswrapper[4715]: I1203 23:17:28.121259 4715 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:28 crc kubenswrapper[4715]: I1203 23:17:28.634293 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:17:28 crc kubenswrapper[4715]: E1203 23:17:28.634804 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:17:29 crc kubenswrapper[4715]: I1203 23:17:29.066465 4715 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wtjn8"] Dec 03 23:17:29 crc kubenswrapper[4715]: I1203 23:17:29.426447 4715 generic.go:334] "Generic (PLEG): container finished" podID="1c39aac8-08c3-4c09-a5ce-86ef40dc357e" containerID="13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e" exitCode=0 Dec 03 23:17:29 crc kubenswrapper[4715]: I1203 23:17:29.426614 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjn8" event={"ID":"1c39aac8-08c3-4c09-a5ce-86ef40dc357e","Type":"ContainerDied","Data":"13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e"} Dec 03 23:17:29 crc kubenswrapper[4715]: I1203 23:17:29.427055 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjn8" event={"ID":"1c39aac8-08c3-4c09-a5ce-86ef40dc357e","Type":"ContainerStarted","Data":"f7f15e1d219ef73f33cabde6c0ffa6c8adcbfac9fb6098c06953e72d95ad29a7"} Dec 03 23:17:29 crc kubenswrapper[4715]: I1203 23:17:29.430272 4715 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 23:17:30 crc kubenswrapper[4715]: I1203 23:17:30.447394 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjn8" event={"ID":"1c39aac8-08c3-4c09-a5ce-86ef40dc357e","Type":"ContainerStarted","Data":"e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd"} Dec 03 23:17:31 crc kubenswrapper[4715]: I1203 23:17:31.460526 4715 generic.go:334] "Generic (PLEG): container finished" podID="1c39aac8-08c3-4c09-a5ce-86ef40dc357e" containerID="e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd" exitCode=0 Dec 03 23:17:31 crc kubenswrapper[4715]: I1203 23:17:31.460609 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjn8" event={"ID":"1c39aac8-08c3-4c09-a5ce-86ef40dc357e","Type":"ContainerDied","Data":"e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd"} Dec 03 23:17:32 crc kubenswrapper[4715]: I1203 23:17:32.477491 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjn8" event={"ID":"1c39aac8-08c3-4c09-a5ce-86ef40dc357e","Type":"ContainerStarted","Data":"e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162"} Dec 03 23:17:32 crc kubenswrapper[4715]: I1203 23:17:32.498027 4715 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wtjn8" podStartSLOduration=3.070976725 podStartE2EDuration="5.498004418s" podCreationTimestamp="2025-12-03 23:17:27 +0000 UTC" firstStartedPulling="2025-12-03 23:17:29.429694764 +0000 UTC m=+5746.172405409" lastFinishedPulling="2025-12-03 23:17:31.856722477 +0000 UTC m=+5748.599433102" observedRunningTime="2025-12-03 23:17:32.497949296 +0000 UTC m=+5749.240659901" watchObservedRunningTime="2025-12-03 23:17:32.498004418 +0000 UTC m=+5749.240715023" Dec 03 23:17:38 crc kubenswrapper[4715]: I1203 23:17:38.122273 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:38 crc kubenswrapper[4715]: I1203 23:17:38.122965 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:38 crc kubenswrapper[4715]: I1203 23:17:38.198693 4715 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:38 crc kubenswrapper[4715]: I1203 23:17:38.592343 4715 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:38 crc kubenswrapper[4715]: I1203 23:17:38.650186 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wtjn8"] Dec 03 23:17:40 crc kubenswrapper[4715]: I1203 23:17:40.565277 4715 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wtjn8" podUID="1c39aac8-08c3-4c09-a5ce-86ef40dc357e" containerName="registry-server" containerID="cri-o://e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162" gracePeriod=2 Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.081292 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.187449 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-catalog-content\") pod \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.187862 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kpvm\" (UniqueName: \"kubernetes.io/projected/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-kube-api-access-2kpvm\") pod \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.187964 4715 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-utilities\") pod \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\" (UID: \"1c39aac8-08c3-4c09-a5ce-86ef40dc357e\") " Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.189751 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-utilities" (OuterVolumeSpecName: "utilities") pod "1c39aac8-08c3-4c09-a5ce-86ef40dc357e" (UID: "1c39aac8-08c3-4c09-a5ce-86ef40dc357e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.194933 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-kube-api-access-2kpvm" (OuterVolumeSpecName: "kube-api-access-2kpvm") pod "1c39aac8-08c3-4c09-a5ce-86ef40dc357e" (UID: "1c39aac8-08c3-4c09-a5ce-86ef40dc357e"). InnerVolumeSpecName "kube-api-access-2kpvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.290176 4715 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kpvm\" (UniqueName: \"kubernetes.io/projected/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-kube-api-access-2kpvm\") on node \"crc\" DevicePath \"\"" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.290221 4715 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.579748 4715 generic.go:334] "Generic (PLEG): container finished" podID="1c39aac8-08c3-4c09-a5ce-86ef40dc357e" containerID="e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162" exitCode=0 Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.579820 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjn8" event={"ID":"1c39aac8-08c3-4c09-a5ce-86ef40dc357e","Type":"ContainerDied","Data":"e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162"} Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.579862 4715 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wtjn8" event={"ID":"1c39aac8-08c3-4c09-a5ce-86ef40dc357e","Type":"ContainerDied","Data":"f7f15e1d219ef73f33cabde6c0ffa6c8adcbfac9fb6098c06953e72d95ad29a7"} Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.579892 4715 scope.go:117] "RemoveContainer" containerID="e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.579909 4715 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wtjn8" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.610751 4715 scope.go:117] "RemoveContainer" containerID="e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.633600 4715 scope.go:117] "RemoveContainer" containerID="13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.634610 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:17:41 crc kubenswrapper[4715]: E1203 23:17:41.634859 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.717467 4715 scope.go:117] "RemoveContainer" containerID="e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162" Dec 03 23:17:41 crc kubenswrapper[4715]: E1203 23:17:41.718127 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162\": container with ID starting with e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162 not found: ID does not exist" containerID="e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.718188 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162"} err="failed to get container status \"e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162\": rpc error: code = NotFound desc = could not find container \"e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162\": container with ID starting with e50610a2e5ef008ef9314a8cff5730e3e340f07d4f8a4aaa698430f477d4e162 not found: ID does not exist" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.718222 4715 scope.go:117] "RemoveContainer" containerID="e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd" Dec 03 23:17:41 crc kubenswrapper[4715]: E1203 23:17:41.718640 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd\": container with ID starting with e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd not found: ID does not exist" containerID="e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.718705 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd"} err="failed to get container status \"e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd\": rpc error: code = NotFound desc = could not find container \"e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd\": container with ID starting with e2b3d80805dc5ce7947c0dd51689ae43aefcb1b0ee81617e21f06c2c044924fd not found: ID does not exist" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.718746 4715 scope.go:117] "RemoveContainer" containerID="13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e" Dec 03 23:17:41 crc kubenswrapper[4715]: E1203 23:17:41.719083 4715 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e\": container with ID starting with 13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e not found: ID does not exist" containerID="13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e" Dec 03 23:17:41 crc kubenswrapper[4715]: I1203 23:17:41.719138 4715 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e"} err="failed to get container status \"13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e\": rpc error: code = NotFound desc = could not find container \"13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e\": container with ID starting with 13f1cfa5f31004c38fa4f41269dfd296d9f74f5e8c0e42d76eae6da92fdffc2e not found: ID does not exist" Dec 03 23:17:42 crc kubenswrapper[4715]: I1203 23:17:42.375863 4715 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c39aac8-08c3-4c09-a5ce-86ef40dc357e" (UID: "1c39aac8-08c3-4c09-a5ce-86ef40dc357e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 23:17:42 crc kubenswrapper[4715]: I1203 23:17:42.412458 4715 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c39aac8-08c3-4c09-a5ce-86ef40dc357e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 23:17:42 crc kubenswrapper[4715]: I1203 23:17:42.514943 4715 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wtjn8"] Dec 03 23:17:42 crc kubenswrapper[4715]: I1203 23:17:42.524391 4715 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wtjn8"] Dec 03 23:17:43 crc kubenswrapper[4715]: I1203 23:17:43.652565 4715 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c39aac8-08c3-4c09-a5ce-86ef40dc357e" path="/var/lib/kubelet/pods/1c39aac8-08c3-4c09-a5ce-86ef40dc357e/volumes" Dec 03 23:17:55 crc kubenswrapper[4715]: I1203 23:17:55.634741 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:17:55 crc kubenswrapper[4715]: E1203 23:17:55.635755 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:18:06 crc kubenswrapper[4715]: I1203 23:18:06.635607 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:18:06 crc kubenswrapper[4715]: E1203 23:18:06.636385 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:18:21 crc kubenswrapper[4715]: I1203 23:18:21.649223 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:18:21 crc kubenswrapper[4715]: E1203 23:18:21.650480 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:18:32 crc kubenswrapper[4715]: I1203 23:18:32.634932 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:18:32 crc kubenswrapper[4715]: E1203 23:18:32.636115 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:18:46 crc kubenswrapper[4715]: I1203 23:18:46.635238 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:18:46 crc kubenswrapper[4715]: E1203 23:18:46.636582 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" Dec 03 23:18:57 crc kubenswrapper[4715]: I1203 23:18:57.635140 4715 scope.go:117] "RemoveContainer" containerID="ac58b5c8453b513d145d2420ddb885e85d809255a04eac6fd4ad00d343924a15" Dec 03 23:18:57 crc kubenswrapper[4715]: E1203 23:18:57.636221 4715 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-ctls2_openshift-machine-config-operator(0ca0ce89-c3ec-421f-9954-c594760b16a9)\"" pod="openshift-machine-config-operator/machine-config-daemon-ctls2" podUID="0ca0ce89-c3ec-421f-9954-c594760b16a9" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114142646024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114142704017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114126515016506 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114126515015456 5ustar corecore